Mar 18 10:12:39 crc systemd[1]: Starting Kubernetes Kubelet... Mar 18 10:12:39 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:39 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 10:12:40 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 18 10:12:40 crc kubenswrapper[4733]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.946286 4733 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949066 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949084 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949089 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949093 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949097 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949101 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949105 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949109 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949113 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949116 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949120 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949124 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949134 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949137 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949141 4733 feature_gate.go:330] unrecognized feature gate: Example Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949146 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949150 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949154 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949158 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949164 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949169 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949173 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949179 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949204 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949208 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949214 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949218 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949223 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949226 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949230 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949234 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949237 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949241 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949246 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949250 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949254 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949258 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949262 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949267 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949272 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949277 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949281 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949286 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949290 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949294 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949299 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949304 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949308 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949312 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949315 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949319 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949322 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949326 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949330 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949334 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949337 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949343 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949347 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949352 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949355 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949359 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949363 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949366 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949370 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949373 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949377 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949380 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949384 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949387 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949390 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.949395 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950707 4733 flags.go:64] FLAG: --address="0.0.0.0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950720 4733 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950729 4733 flags.go:64] FLAG: --anonymous-auth="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950735 4733 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950740 4733 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950744 4733 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950751 4733 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950756 4733 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950761 4733 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950765 4733 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950770 4733 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950774 4733 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950779 4733 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950783 4733 flags.go:64] FLAG: --cgroup-root="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950788 4733 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950792 4733 flags.go:64] FLAG: --client-ca-file="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950796 4733 flags.go:64] FLAG: --cloud-config="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950800 4733 flags.go:64] FLAG: --cloud-provider="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950804 4733 flags.go:64] FLAG: --cluster-dns="[]" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950809 4733 flags.go:64] FLAG: --cluster-domain="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950813 4733 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950817 4733 flags.go:64] FLAG: --config-dir="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950821 4733 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950826 4733 flags.go:64] FLAG: --container-log-max-files="5" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950832 4733 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950836 4733 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950840 4733 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950844 4733 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950848 4733 flags.go:64] FLAG: --contention-profiling="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950854 4733 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950858 4733 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950863 4733 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950866 4733 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950872 4733 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950877 4733 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950882 4733 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950886 4733 flags.go:64] FLAG: --enable-load-reader="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950891 4733 flags.go:64] FLAG: --enable-server="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950895 4733 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950901 4733 flags.go:64] FLAG: --event-burst="100" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950905 4733 flags.go:64] FLAG: --event-qps="50" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950909 4733 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950913 4733 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950917 4733 flags.go:64] FLAG: --eviction-hard="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950922 4733 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950927 4733 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950931 4733 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950936 4733 flags.go:64] FLAG: --eviction-soft="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950940 4733 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950944 4733 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950948 4733 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950952 4733 flags.go:64] FLAG: --experimental-mounter-path="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950956 4733 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950960 4733 flags.go:64] FLAG: --fail-swap-on="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950965 4733 flags.go:64] FLAG: --feature-gates="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950970 4733 flags.go:64] FLAG: --file-check-frequency="20s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950974 4733 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950979 4733 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950983 4733 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950987 4733 flags.go:64] FLAG: --healthz-port="10248" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950991 4733 flags.go:64] FLAG: --help="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.950995 4733 flags.go:64] FLAG: --hostname-override="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951000 4733 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951004 4733 flags.go:64] FLAG: --http-check-frequency="20s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951008 4733 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951012 4733 flags.go:64] FLAG: --image-credential-provider-config="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951016 4733 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951020 4733 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951024 4733 flags.go:64] FLAG: --image-service-endpoint="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951028 4733 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951032 4733 flags.go:64] FLAG: --kube-api-burst="100" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951037 4733 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951041 4733 flags.go:64] FLAG: --kube-api-qps="50" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951045 4733 flags.go:64] FLAG: --kube-reserved="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951049 4733 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951053 4733 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951058 4733 flags.go:64] FLAG: --kubelet-cgroups="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951062 4733 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951068 4733 flags.go:64] FLAG: --lock-file="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951073 4733 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951078 4733 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951082 4733 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951090 4733 flags.go:64] FLAG: --log-json-split-stream="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951096 4733 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951101 4733 flags.go:64] FLAG: --log-text-split-stream="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951105 4733 flags.go:64] FLAG: --logging-format="text" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951109 4733 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951114 4733 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951118 4733 flags.go:64] FLAG: --manifest-url="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951123 4733 flags.go:64] FLAG: --manifest-url-header="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951129 4733 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951133 4733 flags.go:64] FLAG: --max-open-files="1000000" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951138 4733 flags.go:64] FLAG: --max-pods="110" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951142 4733 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951146 4733 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951151 4733 flags.go:64] FLAG: --memory-manager-policy="None" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951155 4733 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951158 4733 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951165 4733 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951169 4733 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951179 4733 flags.go:64] FLAG: --node-status-max-images="50" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951197 4733 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951201 4733 flags.go:64] FLAG: --oom-score-adj="-999" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951205 4733 flags.go:64] FLAG: --pod-cidr="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951209 4733 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951217 4733 flags.go:64] FLAG: --pod-manifest-path="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951221 4733 flags.go:64] FLAG: --pod-max-pids="-1" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951226 4733 flags.go:64] FLAG: --pods-per-core="0" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951230 4733 flags.go:64] FLAG: --port="10250" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951234 4733 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951239 4733 flags.go:64] FLAG: --provider-id="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951243 4733 flags.go:64] FLAG: --qos-reserved="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951247 4733 flags.go:64] FLAG: --read-only-port="10255" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951252 4733 flags.go:64] FLAG: --register-node="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951256 4733 flags.go:64] FLAG: --register-schedulable="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951260 4733 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951268 4733 flags.go:64] FLAG: --registry-burst="10" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951272 4733 flags.go:64] FLAG: --registry-qps="5" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951276 4733 flags.go:64] FLAG: --reserved-cpus="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951280 4733 flags.go:64] FLAG: --reserved-memory="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951286 4733 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951290 4733 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951294 4733 flags.go:64] FLAG: --rotate-certificates="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951298 4733 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951302 4733 flags.go:64] FLAG: --runonce="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951306 4733 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951311 4733 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951315 4733 flags.go:64] FLAG: --seccomp-default="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951319 4733 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951323 4733 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951327 4733 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951332 4733 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951336 4733 flags.go:64] FLAG: --storage-driver-password="root" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951340 4733 flags.go:64] FLAG: --storage-driver-secure="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951344 4733 flags.go:64] FLAG: --storage-driver-table="stats" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951349 4733 flags.go:64] FLAG: --storage-driver-user="root" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951352 4733 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951357 4733 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951361 4733 flags.go:64] FLAG: --system-cgroups="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951365 4733 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951371 4733 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951375 4733 flags.go:64] FLAG: --tls-cert-file="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951385 4733 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951390 4733 flags.go:64] FLAG: --tls-min-version="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951395 4733 flags.go:64] FLAG: --tls-private-key-file="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951399 4733 flags.go:64] FLAG: --topology-manager-policy="none" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951403 4733 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951407 4733 flags.go:64] FLAG: --topology-manager-scope="container" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951412 4733 flags.go:64] FLAG: --v="2" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951418 4733 flags.go:64] FLAG: --version="false" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951423 4733 flags.go:64] FLAG: --vmodule="" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951428 4733 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951433 4733 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951530 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951536 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951540 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951544 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951548 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951552 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951556 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951561 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951565 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951568 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951572 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951576 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951579 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951583 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951586 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951590 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951594 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951597 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951602 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951606 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951611 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951616 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951620 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951624 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951627 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951632 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951636 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951639 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951644 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951648 4733 feature_gate.go:330] unrecognized feature gate: Example Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951652 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951656 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951660 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951664 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951667 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951671 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951675 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951678 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951682 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951686 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951689 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951693 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951697 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951700 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951704 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951707 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951711 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951714 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951718 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951722 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951725 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951729 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951733 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951737 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951741 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951745 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951749 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951752 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951756 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951759 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951763 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951767 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951771 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951775 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951778 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951782 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951785 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951789 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951793 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951796 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.951800 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.951806 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.961714 4733 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.962162 4733 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962296 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962307 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962313 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962320 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962326 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962332 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962337 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962342 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962348 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962354 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962359 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962364 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962369 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962374 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962380 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962385 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962390 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962395 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962401 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962406 4733 feature_gate.go:330] unrecognized feature gate: Example Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962411 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962416 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962422 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962427 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962433 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962438 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962443 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962448 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962453 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962459 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962465 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962472 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962478 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962483 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962490 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962496 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962501 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962507 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962512 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962517 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962523 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962528 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962533 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962541 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962550 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962556 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962563 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962570 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962577 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962583 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962588 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962595 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962600 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962605 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962610 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962616 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962621 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962627 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962632 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962638 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962643 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962648 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962653 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962662 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962669 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962674 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962680 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962685 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962690 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962696 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962702 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.962711 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962860 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962870 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962876 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962882 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962887 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962893 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962898 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962903 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962909 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962916 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962924 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962931 4733 feature_gate.go:330] unrecognized feature gate: Example Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962938 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962944 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962950 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962956 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962964 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962969 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962975 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962981 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962986 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962991 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.962996 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963002 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963008 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963013 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963018 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963024 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963029 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963034 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963041 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963048 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963054 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963059 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963066 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963071 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963076 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963082 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963087 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963092 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963098 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963103 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963108 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963113 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963119 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963125 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963131 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963136 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963144 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963151 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963157 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963162 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963167 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963173 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963178 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963200 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963206 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963212 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963217 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963222 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963227 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963233 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963238 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963243 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963248 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963253 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963259 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963265 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963270 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963275 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 10:12:40 crc kubenswrapper[4733]: W0318 10:12:40.963281 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.963292 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.964369 4733 server.go:940] "Client rotation is on, will bootstrap in background" Mar 18 10:12:40 crc kubenswrapper[4733]: E0318 10:12:40.971934 4733 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.977084 4733 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.977211 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.978881 4733 server.go:997] "Starting client certificate rotation" Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.978915 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 18 10:12:40 crc kubenswrapper[4733]: I0318 10:12:40.979124 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.004512 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.006995 4733 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.010351 4733 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.027747 4733 log.go:25] "Validated CRI v1 runtime API" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.068970 4733 log.go:25] "Validated CRI v1 image API" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.073898 4733 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.079336 4733 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-18-10-08-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.079376 4733 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.095542 4733 manager.go:217] Machine: {Timestamp:2026-03-18 10:12:41.094050248 +0000 UTC m=+0.585784593 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fe704b25-4cdf-410a-9afb-ebc7963f4bc5 BootID:5a826494-c246-4717-869b-fd136e2b8410 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:07:57:e6 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:07:57:e6 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f8:93:62 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a7:c6:3d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1c:3f:12 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:fc:7f:71 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:66:97:9e:22:cf:16 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:d7:ed:05:64:96 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.095769 4733 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.095964 4733 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.097027 4733 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.097197 4733 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.097229 4733 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.097400 4733 topology_manager.go:138] "Creating topology manager with none policy" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.097410 4733 container_manager_linux.go:303] "Creating device plugin manager" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.098151 4733 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.098943 4733 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.099154 4733 state_mem.go:36] "Initialized new in-memory state store" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.099577 4733 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.106130 4733 kubelet.go:418] "Attempting to sync node with API server" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.106149 4733 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.106198 4733 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.106211 4733 kubelet.go:324] "Adding apiserver pod source" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.106221 4733 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.111392 4733 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.112349 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.114886 4733 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.116206 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.116285 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.116295 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.116334 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116439 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116464 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116473 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116483 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116496 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116506 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116518 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116531 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116541 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116550 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116582 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.116592 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.117653 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.118113 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.118210 4733 server.go:1280] "Started kubelet" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.118334 4733 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.118853 4733 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.119030 4733 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 18 10:12:41 crc systemd[1]: Started Kubernetes Kubelet. Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.120553 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.120584 4733 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.120830 4733 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.120853 4733 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.120854 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.120894 4733 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.121403 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.121460 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.121899 4733 factory.go:55] Registering systemd factory Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.122989 4733 factory.go:221] Registration of the systemd container factory successfully Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.125067 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.125460 4733 factory.go:153] Registering CRI-O factory Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.125493 4733 factory.go:221] Registration of the crio container factory successfully Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.125612 4733 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.126111 4733 factory.go:103] Registering Raw factory Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.126167 4733 manager.go:1196] Started watching for new ooms in manager Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.127062 4733 server.go:460] "Adding debug handlers to kubelet server" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.127109 4733 manager.go:319] Starting recovery of all containers Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.130746 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189de7dc2ed1cc5e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,LastTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135080 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135138 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135152 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135166 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135178 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135217 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135235 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135249 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135265 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135276 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135314 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135326 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135340 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135353 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135366 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135378 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135391 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135404 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135417 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135428 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135439 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135451 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135463 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135475 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135487 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135500 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135514 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135528 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135541 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135553 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135567 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135580 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135592 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135626 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135640 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135655 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135667 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135679 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135692 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135704 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135716 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135729 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135741 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135754 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135765 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135777 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135791 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135802 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135814 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135826 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135836 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135849 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135866 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135881 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135895 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135907 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135921 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135934 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135946 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135960 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135972 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135985 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.135999 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136012 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136028 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136039 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136050 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136062 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136075 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136086 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136099 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136113 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136125 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136137 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136152 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136164 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136177 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136220 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136239 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136252 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136269 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136286 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136334 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136350 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136365 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136381 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136395 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136415 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136431 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136447 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136465 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136482 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136498 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136513 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136531 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136546 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136568 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136583 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136599 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136614 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136628 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136643 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136658 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136673 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136697 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136716 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136732 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136747 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136765 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136783 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136802 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.136822 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139781 4733 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139835 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139858 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139877 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139902 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139918 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139931 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139944 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139957 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139970 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139982 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.139996 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140009 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140022 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140075 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140092 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140121 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140137 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140152 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140164 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140177 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140207 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140226 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140263 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140278 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140322 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140337 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140352 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140364 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140377 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140390 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140403 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140416 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140429 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140442 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140457 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140469 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140492 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140506 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140519 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140532 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140545 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140559 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140571 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140586 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140600 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140614 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140628 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140641 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140654 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140667 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140681 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140694 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140710 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140736 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140753 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140767 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140778 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140792 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140804 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140818 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140831 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140843 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140856 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140869 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140880 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140893 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140905 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140920 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140936 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140952 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140964 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140975 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140987 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.140999 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141012 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141025 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141038 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141052 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141064 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141077 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141089 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141102 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141115 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141128 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141139 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141154 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141167 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141179 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141213 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141227 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141240 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141255 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141269 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141282 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141296 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141310 4733 reconstruct.go:97] "Volume reconstruction finished" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.141319 4733 reconciler.go:26] "Reconciler: start to sync state" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.148201 4733 manager.go:324] Recovery completed Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.156906 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.158754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.158789 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.158799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.159850 4733 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.159873 4733 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.159927 4733 state_mem.go:36] "Initialized new in-memory state store" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.168691 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.170303 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.174176 4733 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.174258 4733 kubelet.go:2335] "Starting kubelet main sync loop" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.174311 4733 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.178373 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.178454 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.178739 4733 policy_none.go:49] "None policy: Start" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.179607 4733 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.179642 4733 state_mem.go:35] "Initializing new in-memory state store" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.221839 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.238431 4733 manager.go:334] "Starting Device Plugin manager" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.238506 4733 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.238526 4733 server.go:79] "Starting device plugin registration server" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.239126 4733 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.239151 4733 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.239393 4733 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.239503 4733 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.239513 4733 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.246976 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.275122 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.275293 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276496 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.276927 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277348 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277509 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277568 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277975 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.277992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.278153 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.278320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.278348 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279382 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279595 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279701 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.279765 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.280283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.280315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.280330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.281291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.281348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.281369 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.281709 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.281770 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.282238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.282277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.282299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.283069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.283115 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.283127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.325851 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.340051 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.341230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.341283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.341302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.341336 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.341745 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343894 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343937 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343965 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.343987 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344023 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344082 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344264 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344344 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344423 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344471 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.344509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445654 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445768 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445795 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445822 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445845 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445873 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445893 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445931 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445946 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445967 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445929 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446006 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446040 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446010 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445968 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.445959 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446144 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446209 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446296 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446328 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446315 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446374 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446378 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.446423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.542695 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.544365 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.544423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.544437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.544469 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.545113 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.612003 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.618656 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.641349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.648210 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.651897 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.663247 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7c4e20ab044922246caebfc659a304aa49b24d1ccc341f550cd8a80d605b70e3 WatchSource:0}: Error finding container 7c4e20ab044922246caebfc659a304aa49b24d1ccc341f550cd8a80d605b70e3: Status 404 returned error can't find the container with id 7c4e20ab044922246caebfc659a304aa49b24d1ccc341f550cd8a80d605b70e3 Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.665541 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-38a5e4c60bea1129487e99ba2f872c1fb6c8391243dbca310e4a331c62769fa0 WatchSource:0}: Error finding container 38a5e4c60bea1129487e99ba2f872c1fb6c8391243dbca310e4a331c62769fa0: Status 404 returned error can't find the container with id 38a5e4c60bea1129487e99ba2f872c1fb6c8391243dbca310e4a331c62769fa0 Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.672468 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5c6b08f6b5ce097065e8ca2e22e00fdd77fd84c37944d0dabf315a6cecdf4a94 WatchSource:0}: Error finding container 5c6b08f6b5ce097065e8ca2e22e00fdd77fd84c37944d0dabf315a6cecdf4a94: Status 404 returned error can't find the container with id 5c6b08f6b5ce097065e8ca2e22e00fdd77fd84c37944d0dabf315a6cecdf4a94 Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.673959 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d997e6f39090acbf8aab31a8d6d3ea67c29b8216bbc731bc93b42649efceae2d WatchSource:0}: Error finding container d997e6f39090acbf8aab31a8d6d3ea67c29b8216bbc731bc93b42649efceae2d: Status 404 returned error can't find the container with id d997e6f39090acbf8aab31a8d6d3ea67c29b8216bbc731bc93b42649efceae2d Mar 18 10:12:41 crc kubenswrapper[4733]: W0318 10:12:41.678180 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e2784b6b4caac6f5512efe62b0fe10e9443790c72530dc26d4664e50fd4a3cbf WatchSource:0}: Error finding container e2784b6b4caac6f5512efe62b0fe10e9443790c72530dc26d4664e50fd4a3cbf: Status 404 returned error can't find the container with id e2784b6b4caac6f5512efe62b0fe10e9443790c72530dc26d4664e50fd4a3cbf Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.726995 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.946028 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.947786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.947825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.947838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:41 crc kubenswrapper[4733]: I0318 10:12:41.947864 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:41 crc kubenswrapper[4733]: E0318 10:12:41.948330 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.119597 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:42 crc kubenswrapper[4733]: W0318 10:12:42.126403 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.126483 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.178554 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e2784b6b4caac6f5512efe62b0fe10e9443790c72530dc26d4664e50fd4a3cbf"} Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.180057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d997e6f39090acbf8aab31a8d6d3ea67c29b8216bbc731bc93b42649efceae2d"} Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.181122 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5c6b08f6b5ce097065e8ca2e22e00fdd77fd84c37944d0dabf315a6cecdf4a94"} Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.182918 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"38a5e4c60bea1129487e99ba2f872c1fb6c8391243dbca310e4a331c62769fa0"} Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.183934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7c4e20ab044922246caebfc659a304aa49b24d1ccc341f550cd8a80d605b70e3"} Mar 18 10:12:42 crc kubenswrapper[4733]: W0318 10:12:42.526086 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.526270 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.528571 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Mar 18 10:12:42 crc kubenswrapper[4733]: W0318 10:12:42.599525 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.599623 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:42 crc kubenswrapper[4733]: W0318 10:12:42.662818 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.662897 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.662970 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189de7dc2ed1cc5e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,LastTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.749415 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.751297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.751342 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.751371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:42 crc kubenswrapper[4733]: I0318 10:12:42.751402 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:42 crc kubenswrapper[4733]: E0318 10:12:42.751956 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.119240 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.189369 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.189412 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.189425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.189434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.189580 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.190824 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b" exitCode=0 Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.190854 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.190990 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.191122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.191144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.191154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.192672 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.192739 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.192779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.192790 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.193353 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8" exitCode=0 Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.193446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.193562 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: E0318 10:12:43.198065 4733 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.198742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.198829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.198908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.200437 4733 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03" exitCode=0 Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.200627 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.200680 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.201917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.202025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.202107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.202378 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734" exitCode=0 Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.202425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734"} Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.202581 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.204956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.205971 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.206071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.215372 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.216496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.216670 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:43 crc kubenswrapper[4733]: I0318 10:12:43.216692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.118720 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:44 crc kubenswrapper[4733]: E0318 10:12:44.129683 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="3.2s" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.207372 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d" exitCode=0 Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.207459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.207504 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.208713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.208748 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.208759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.211920 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.211938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.216583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.216621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.216632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.220314 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.220365 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.220377 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.221835 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.223459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.223517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.223534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.225640 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.225977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226027 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226041 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226051 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba"} Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226881 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.226947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: W0318 10:12:44.312705 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:44 crc kubenswrapper[4733]: E0318 10:12:44.312778 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.352803 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.354564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.354609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.354621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:44 crc kubenswrapper[4733]: I0318 10:12:44.354647 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:44 crc kubenswrapper[4733]: E0318 10:12:44.355209 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.184:6443: connect: connection refused" node="crc" Mar 18 10:12:44 crc kubenswrapper[4733]: W0318 10:12:44.457797 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.184:6443: connect: connection refused Mar 18 10:12:44 crc kubenswrapper[4733]: E0318 10:12:44.457882 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.184:6443: connect: connection refused" logger="UnhandledError" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.231705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"87f497bc1bba8e258496f93af9e62eef320f9f59e6180ea36c438694242afd46"} Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.231846 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.232703 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.232735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.232747 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.234691 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5" exitCode=0 Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.234800 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.235396 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.235846 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5"} Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.235891 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.235947 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.236623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.236649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.236661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237167 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237222 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:45 crc kubenswrapper[4733]: I0318 10:12:45.237717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243469 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432"} Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243526 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420"} Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df"} Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243558 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4"} Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243570 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556"} Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243569 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243642 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243669 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.243684 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245161 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245173 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.245941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.514785 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.643963 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.644360 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.646071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.646118 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:46 crc kubenswrapper[4733]: I0318 10:12:46.646131 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.248156 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.248156 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.250270 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.290493 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.555649 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.557252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.557313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.557331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.557365 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:12:47 crc kubenswrapper[4733]: I0318 10:12:47.708269 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.251386 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.253234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.253306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.253327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.708472 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.708736 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.710805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.710879 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.710903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.889305 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.897671 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:48 crc kubenswrapper[4733]: I0318 10:12:48.916074 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.047866 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.048069 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.049762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.049797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.049806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.254967 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.254967 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.256317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.256375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.256393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.257068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.257117 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:49 crc kubenswrapper[4733]: I0318 10:12:49.257136 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.158347 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.158650 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.160380 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.160469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.160491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.257942 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.259548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.259625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:50 crc kubenswrapper[4733]: I0318 10:12:50.259647 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:51 crc kubenswrapper[4733]: E0318 10:12:51.247309 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:12:51 crc kubenswrapper[4733]: I0318 10:12:51.782727 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:51 crc kubenswrapper[4733]: I0318 10:12:51.782947 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:51 crc kubenswrapper[4733]: I0318 10:12:51.784096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:51 crc kubenswrapper[4733]: I0318 10:12:51.784145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:51 crc kubenswrapper[4733]: I0318 10:12:51.784162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:54 crc kubenswrapper[4733]: I0318 10:12:54.783336 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:12:54 crc kubenswrapper[4733]: I0318 10:12:54.783454 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:12:54 crc kubenswrapper[4733]: W0318 10:12:54.793069 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 10:12:54 crc kubenswrapper[4733]: I0318 10:12:54.793548 4733 trace.go:236] Trace[308787041]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 10:12:44.792) (total time: 10001ms): Mar 18 10:12:54 crc kubenswrapper[4733]: Trace[308787041]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (10:12:54.793) Mar 18 10:12:54 crc kubenswrapper[4733]: Trace[308787041]: [10.001445119s] [10.001445119s] END Mar 18 10:12:54 crc kubenswrapper[4733]: E0318 10:12:54.793850 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 10:12:54 crc kubenswrapper[4733]: E0318 10:12:54.998528 4733 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:12:54 crc kubenswrapper[4733]: E0318 10:12:54.999598 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:54Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 10:12:55 crc kubenswrapper[4733]: E0318 10:12:55.002787 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.003605 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z Mar 18 10:12:55 crc kubenswrapper[4733]: W0318 10:12:55.003636 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z Mar 18 10:12:55 crc kubenswrapper[4733]: E0318 10:12:55.003713 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:12:55 crc kubenswrapper[4733]: W0318 10:12:55.005554 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z Mar 18 10:12:55 crc kubenswrapper[4733]: E0318 10:12:55.005622 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.007291 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.007543 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 10:12:55 crc kubenswrapper[4733]: W0318 10:12:55.009181 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z Mar 18 10:12:55 crc kubenswrapper[4733]: E0318 10:12:55.009254 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:12:55 crc kubenswrapper[4733]: E0318 10:12:55.012081 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189de7dc2ed1cc5e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,LastTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.018529 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.018826 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.122498 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:55Z is after 2026-02-23T05:33:13Z Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.272677 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.276662 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="87f497bc1bba8e258496f93af9e62eef320f9f59e6180ea36c438694242afd46" exitCode=255 Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.276717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"87f497bc1bba8e258496f93af9e62eef320f9f59e6180ea36c438694242afd46"} Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.276906 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.277770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.277798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.277808 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:55 crc kubenswrapper[4733]: I0318 10:12:55.278282 4733 scope.go:117] "RemoveContainer" containerID="87f497bc1bba8e258496f93af9e62eef320f9f59e6180ea36c438694242afd46" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.122593 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:56Z is after 2026-02-23T05:33:13Z Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.283297 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.285932 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95"} Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.286100 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.287044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.287096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.287110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.515583 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.648915 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.649097 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.650822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.650886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:56 crc kubenswrapper[4733]: I0318 10:12:56.650914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.125041 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:57Z is after 2026-02-23T05:33:13Z Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.291181 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.291994 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.294587 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" exitCode=255 Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.294745 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95"} Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.294881 4733 scope.go:117] "RemoveContainer" containerID="87f497bc1bba8e258496f93af9e62eef320f9f59e6180ea36c438694242afd46" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.295012 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.296014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.296058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.296071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:57 crc kubenswrapper[4733]: I0318 10:12:57.296741 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:12:57 crc kubenswrapper[4733]: E0318 10:12:57.296934 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.125477 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:58Z is after 2026-02-23T05:33:13Z Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.299432 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.301809 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.302750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.302793 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.302806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.303468 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:12:58 crc kubenswrapper[4733]: E0318 10:12:58.303643 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.316386 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:58 crc kubenswrapper[4733]: I0318 10:12:58.925160 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.086755 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.086994 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.088504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.088576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.088595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.105438 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.121972 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:59Z is after 2026-02-23T05:33:13Z Mar 18 10:12:59 crc kubenswrapper[4733]: W0318 10:12:59.214289 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:59Z is after 2026-02-23T05:33:13Z Mar 18 10:12:59 crc kubenswrapper[4733]: E0318 10:12:59.214412 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.304097 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.304132 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305372 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.305548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.306372 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:12:59 crc kubenswrapper[4733]: E0318 10:12:59.306642 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:12:59 crc kubenswrapper[4733]: I0318 10:12:59.310827 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:12:59 crc kubenswrapper[4733]: W0318 10:12:59.384000 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:59Z is after 2026-02-23T05:33:13Z Mar 18 10:12:59 crc kubenswrapper[4733]: E0318 10:12:59.384146 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:12:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.121844 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:13:00Z is after 2026-02-23T05:33:13Z Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.307228 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.308014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.308064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.308076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:00 crc kubenswrapper[4733]: I0318 10:13:00.308790 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:13:00 crc kubenswrapper[4733]: E0318 10:13:00.309010 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.123318 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:01 crc kubenswrapper[4733]: E0318 10:13:01.247483 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.309578 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.310479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.310510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.310520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.311071 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:13:01 crc kubenswrapper[4733]: E0318 10:13:01.311265 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.399761 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.401347 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.401408 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.401431 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:01 crc kubenswrapper[4733]: I0318 10:13:01.401466 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:01 crc kubenswrapper[4733]: E0318 10:13:01.406621 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:01 crc kubenswrapper[4733]: E0318 10:13:01.407626 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:02 crc kubenswrapper[4733]: I0318 10:13:02.122649 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:03 crc kubenswrapper[4733]: I0318 10:13:03.105168 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 10:13:03 crc kubenswrapper[4733]: I0318 10:13:03.127642 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:03 crc kubenswrapper[4733]: I0318 10:13:03.128724 4733 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 10:13:04 crc kubenswrapper[4733]: I0318 10:13:04.125807 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:04 crc kubenswrapper[4733]: I0318 10:13:04.782983 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:13:04 crc kubenswrapper[4733]: I0318 10:13:04.783108 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.018277 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc2ed1cc5e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,LastTimestamp:2026-03-18 10:12:41.11816611 +0000 UTC m=+0.609900435,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.025533 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.030776 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.037633 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.042935 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc36d5f239 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.252655673 +0000 UTC m=+0.744389998,LastTimestamp:2026-03-18 10:12:41.252655673 +0000 UTC m=+0.744389998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.049891 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.276343393 +0000 UTC m=+0.768077718,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.056356 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.276363744 +0000 UTC m=+0.768098059,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.064358 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.276371284 +0000 UTC m=+0.768105609,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.071666 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.277255888 +0000 UTC m=+0.768990213,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.079180 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.277266878 +0000 UTC m=+0.769001203,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.084778 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.277274929 +0000 UTC m=+0.769009244,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.089626 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.277847354 +0000 UTC m=+0.769581679,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.097231 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.277862984 +0000 UTC m=+0.769597309,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.102403 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.277871145 +0000 UTC m=+0.769605470,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.107854 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.277966227 +0000 UTC m=+0.769700572,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.112135 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.277985818 +0000 UTC m=+0.769720153,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.117964 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.278002758 +0000 UTC m=+0.769737093,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: I0318 10:13:05.123144 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.123376 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.279341854 +0000 UTC m=+0.771076219,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.130064 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.279372115 +0000 UTC m=+0.771106480,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.135138 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.279393646 +0000 UTC m=+0.771128011,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.140721 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.279413736 +0000 UTC m=+0.771148081,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.146034 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.279431327 +0000 UTC m=+0.771165662,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.150101 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313de6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313de6ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158805228 +0000 UTC m=+0.650539553,LastTimestamp:2026-03-18 10:12:41.279447647 +0000 UTC m=+0.771181992,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.154457 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313d7e6d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313d7e6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158778477 +0000 UTC m=+0.650512802,LastTimestamp:2026-03-18 10:12:41.28030627 +0000 UTC m=+0.772040605,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.158612 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189de7dc313dbfa0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189de7dc313dbfa0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.158795168 +0000 UTC m=+0.650529493,LastTimestamp:2026-03-18 10:12:41.280324621 +0000 UTC m=+0.772058966,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.163224 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dc4fa94a34 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.669159476 +0000 UTC m=+1.160893821,LastTimestamp:2026-03-18 10:12:41.669159476 +0000 UTC m=+1.160893821,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.169138 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc4fc702f7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.671107319 +0000 UTC m=+1.162841674,LastTimestamp:2026-03-18 10:12:41.671107319 +0000 UTC m=+1.162841674,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.173107 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dc50189b61 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.676454753 +0000 UTC m=+1.168189088,LastTimestamp:2026-03-18 10:12:41.676454753 +0000 UTC m=+1.168189088,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.177586 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dc501a6cfa openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.676573946 +0000 UTC m=+1.168308271,LastTimestamp:2026-03-18 10:12:41.676573946 +0000 UTC m=+1.168308271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.183720 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dc50576aa4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:41.680571044 +0000 UTC m=+1.172305389,LastTimestamp:2026-03-18 10:12:41.680571044 +0000 UTC m=+1.172305389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.188836 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc6fd28db4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.208734644 +0000 UTC m=+1.700468969,LastTimestamp:2026-03-18 10:12:42.208734644 +0000 UTC m=+1.700468969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.192500 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dc70027795 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.211874709 +0000 UTC m=+1.703609034,LastTimestamp:2026-03-18 10:12:42.211874709 +0000 UTC m=+1.703609034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.196504 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dc7009b904 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.212350212 +0000 UTC m=+1.704084537,LastTimestamp:2026-03-18 10:12:42.212350212 +0000 UTC m=+1.704084537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.200277 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dc700f82e4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.212729572 +0000 UTC m=+1.704463927,LastTimestamp:2026-03-18 10:12:42.212729572 +0000 UTC m=+1.704463927,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.204455 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dc7046a93e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.21634387 +0000 UTC m=+1.708078195,LastTimestamp:2026-03-18 10:12:42.21634387 +0000 UTC m=+1.708078195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.208655 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc70a1d2e7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.222318311 +0000 UTC m=+1.714052636,LastTimestamp:2026-03-18 10:12:42.222318311 +0000 UTC m=+1.714052636,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.213732 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc70bd9cce openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.22413947 +0000 UTC m=+1.715873795,LastTimestamp:2026-03-18 10:12:42.22413947 +0000 UTC m=+1.715873795,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.218337 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dc70def0df openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.226323679 +0000 UTC m=+1.718058044,LastTimestamp:2026-03-18 10:12:42.226323679 +0000 UTC m=+1.718058044,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.223219 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dc70f7b1f5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.227945973 +0000 UTC m=+1.719680338,LastTimestamp:2026-03-18 10:12:42.227945973 +0000 UTC m=+1.719680338,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.229884 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dc70fa1237 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.228101687 +0000 UTC m=+1.719836012,LastTimestamp:2026-03-18 10:12:42.228101687 +0000 UTC m=+1.719836012,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.233797 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dc713e36e8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.232567528 +0000 UTC m=+1.724301853,LastTimestamp:2026-03-18 10:12:42.232567528 +0000 UTC m=+1.724301853,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.237782 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc8408c6ec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.547832556 +0000 UTC m=+2.039566881,LastTimestamp:2026-03-18 10:12:42.547832556 +0000 UTC m=+2.039566881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.242262 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc84ee1035 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.562859061 +0000 UTC m=+2.054593386,LastTimestamp:2026-03-18 10:12:42.562859061 +0000 UTC m=+2.054593386,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.248359 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc85031645 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.564236869 +0000 UTC m=+2.055971194,LastTimestamp:2026-03-18 10:12:42.564236869 +0000 UTC m=+2.055971194,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.253515 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc939efa31 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.809334321 +0000 UTC m=+2.301068646,LastTimestamp:2026-03-18 10:12:42.809334321 +0000 UTC m=+2.301068646,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.258067 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc94583a17 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.821474839 +0000 UTC m=+2.313209164,LastTimestamp:2026-03-18 10:12:42.821474839 +0000 UTC m=+2.313209164,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.262398 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc946731d5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.822455765 +0000 UTC m=+2.314190090,LastTimestamp:2026-03-18 10:12:42.822455765 +0000 UTC m=+2.314190090,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.266552 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dca3030a77 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.067550327 +0000 UTC m=+2.559284662,LastTimestamp:2026-03-18 10:12:43.067550327 +0000 UTC m=+2.559284662,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.270391 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dca3f7735c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.083567964 +0000 UTC m=+2.575302309,LastTimestamp:2026-03-18 10:12:43.083567964 +0000 UTC m=+2.575302309,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.275384 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dcaacea3df openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.198333919 +0000 UTC m=+2.690068244,LastTimestamp:2026-03-18 10:12:43.198333919 +0000 UTC m=+2.690068244,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.279518 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dcab02c92a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.201751338 +0000 UTC m=+2.693485703,LastTimestamp:2026-03-18 10:12:43.201751338 +0000 UTC m=+2.693485703,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.283757 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcab24f2c9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.203990217 +0000 UTC m=+2.695724582,LastTimestamp:2026-03-18 10:12:43.203990217 +0000 UTC m=+2.695724582,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.288071 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcabccd2e8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.214992104 +0000 UTC m=+2.706726449,LastTimestamp:2026-03-18 10:12:43.214992104 +0000 UTC m=+2.706726449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.298632 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcb9dcf918 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.45093148 +0000 UTC m=+2.942665805,LastTimestamp:2026-03-18 10:12:43.45093148 +0000 UTC m=+2.942665805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.303414 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dcba060330 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.45362104 +0000 UTC m=+2.945355365,LastTimestamp:2026-03-18 10:12:43.45362104 +0000 UTC m=+2.945355365,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.307159 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcba1167d4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.4543677 +0000 UTC m=+2.946102025,LastTimestamp:2026-03-18 10:12:43.4543677 +0000 UTC m=+2.946102025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.311821 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dcba173e7a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.45475033 +0000 UTC m=+2.946484655,LastTimestamp:2026-03-18 10:12:43.45475033 +0000 UTC m=+2.946484655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.316109 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcbb0c43d3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.470808019 +0000 UTC m=+2.962542344,LastTimestamp:2026-03-18 10:12:43.470808019 +0000 UTC m=+2.962542344,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.321027 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcbb1c428a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.471856266 +0000 UTC m=+2.963590591,LastTimestamp:2026-03-18 10:12:43.471856266 +0000 UTC m=+2.963590591,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.324857 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcbb7127c3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.477419971 +0000 UTC m=+2.969154296,LastTimestamp:2026-03-18 10:12:43.477419971 +0000 UTC m=+2.969154296,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: W0318 10:13:05.325432 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.325520 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.329697 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dcbb8ccb93 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.479231379 +0000 UTC m=+2.970965704,LastTimestamp:2026-03-18 10:12:43.479231379 +0000 UTC m=+2.970965704,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.333903 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189de7dcbc1b1478 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.488556152 +0000 UTC m=+2.980290477,LastTimestamp:2026-03-18 10:12:43.488556152 +0000 UTC m=+2.980290477,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.338955 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcbc405926 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.490998566 +0000 UTC m=+2.982732891,LastTimestamp:2026-03-18 10:12:43.490998566 +0000 UTC m=+2.982732891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.343408 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcc848ac4d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.692870733 +0000 UTC m=+3.184605058,LastTimestamp:2026-03-18 10:12:43.692870733 +0000 UTC m=+3.184605058,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.347426 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcc8746360 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.695735648 +0000 UTC m=+3.187469973,LastTimestamp:2026-03-18 10:12:43.695735648 +0000 UTC m=+3.187469973,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.352495 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcc91a3f00 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.706605312 +0000 UTC m=+3.198339637,LastTimestamp:2026-03-18 10:12:43.706605312 +0000 UTC m=+3.198339637,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.356913 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcc92e773a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.707930426 +0000 UTC m=+3.199664751,LastTimestamp:2026-03-18 10:12:43.707930426 +0000 UTC m=+3.199664751,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.361579 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcc9417af9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.709176569 +0000 UTC m=+3.200910894,LastTimestamp:2026-03-18 10:12:43.709176569 +0000 UTC m=+3.200910894,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.366567 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcc95d7c89 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.711011977 +0000 UTC m=+3.202746302,LastTimestamp:2026-03-18 10:12:43.711011977 +0000 UTC m=+3.202746302,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.372306 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcd5eec0b2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.921858738 +0000 UTC m=+3.413593063,LastTimestamp:2026-03-18 10:12:43.921858738 +0000 UTC m=+3.413593063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.377324 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcd5f08049 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.921973321 +0000 UTC m=+3.413707646,LastTimestamp:2026-03-18 10:12:43.921973321 +0000 UTC m=+3.413707646,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.381810 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcd79b26bf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.949934271 +0000 UTC m=+3.441668596,LastTimestamp:2026-03-18 10:12:43.949934271 +0000 UTC m=+3.441668596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.387124 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcd7ab4543 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.950990659 +0000 UTC m=+3.442724984,LastTimestamp:2026-03-18 10:12:43.950990659 +0000 UTC m=+3.442724984,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.393918 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189de7dcd7dddf13 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:43.954306835 +0000 UTC m=+3.446041160,LastTimestamp:2026-03-18 10:12:43.954306835 +0000 UTC m=+3.446041160,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.399714 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dce03c91ad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.094730669 +0000 UTC m=+3.586465004,LastTimestamp:2026-03-18 10:12:44.094730669 +0000 UTC m=+3.586465004,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.403042 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dce0d14fb4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.104478644 +0000 UTC m=+3.596212969,LastTimestamp:2026-03-18 10:12:44.104478644 +0000 UTC m=+3.596212969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.408419 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dce0dd72da openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.105274074 +0000 UTC m=+3.597008399,LastTimestamp:2026-03-18 10:12:44.105274074 +0000 UTC m=+3.597008399,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.414150 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dce724d1ea openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.210614762 +0000 UTC m=+3.702349087,LastTimestamp:2026-03-18 10:12:44.210614762 +0000 UTC m=+3.702349087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.419465 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dced2cd1a0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.311802272 +0000 UTC m=+3.803536597,LastTimestamp:2026-03-18 10:12:44.311802272 +0000 UTC m=+3.803536597,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.426364 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcee58714f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.331438415 +0000 UTC m=+3.823172740,LastTimestamp:2026-03-18 10:12:44.331438415 +0000 UTC m=+3.823172740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.434441 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dcf36fd294 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.416856724 +0000 UTC m=+3.908591039,LastTimestamp:2026-03-18 10:12:44.416856724 +0000 UTC m=+3.908591039,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.438686 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dcf3f9ef5c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.42590806 +0000 UTC m=+3.917642395,LastTimestamp:2026-03-18 10:12:44.42590806 +0000 UTC m=+3.917642395,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.444001 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd246bf430 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.238686768 +0000 UTC m=+4.730421103,LastTimestamp:2026-03-18 10:12:45.238686768 +0000 UTC m=+4.730421103,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.447178 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd3188d174 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.458682228 +0000 UTC m=+4.950416553,LastTimestamp:2026-03-18 10:12:45.458682228 +0000 UTC m=+4.950416553,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.451760 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd322aebb0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.469305776 +0000 UTC m=+4.961040101,LastTimestamp:2026-03-18 10:12:45.469305776 +0000 UTC m=+4.961040101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.457008 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd323bed29 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.470420265 +0000 UTC m=+4.962154590,LastTimestamp:2026-03-18 10:12:45.470420265 +0000 UTC m=+4.962154590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.462591 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd3d302902 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.65419853 +0000 UTC m=+5.145932845,LastTimestamp:2026-03-18 10:12:45.65419853 +0000 UTC m=+5.145932845,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.468175 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd3de3abaf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.665962927 +0000 UTC m=+5.157697242,LastTimestamp:2026-03-18 10:12:45.665962927 +0000 UTC m=+5.157697242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.474893 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd3df6c92c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.66721566 +0000 UTC m=+5.158949975,LastTimestamp:2026-03-18 10:12:45.66721566 +0000 UTC m=+5.158949975,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.479771 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd47513574 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.824136564 +0000 UTC m=+5.315870899,LastTimestamp:2026-03-18 10:12:45.824136564 +0000 UTC m=+5.315870899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.486361 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd47f875b0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.83509752 +0000 UTC m=+5.326831845,LastTimestamp:2026-03-18 10:12:45.83509752 +0000 UTC m=+5.326831845,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.491378 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd4805baa7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:45.835967143 +0000 UTC m=+5.327701468,LastTimestamp:2026-03-18 10:12:45.835967143 +0000 UTC m=+5.327701468,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.496634 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd522bb822 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:46.006229026 +0000 UTC m=+5.497963361,LastTimestamp:2026-03-18 10:12:46.006229026 +0000 UTC m=+5.497963361,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.503857 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd52afb587 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:46.014879111 +0000 UTC m=+5.506613466,LastTimestamp:2026-03-18 10:12:46.014879111 +0000 UTC m=+5.506613466,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.509369 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd52bd57a9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:46.015772585 +0000 UTC m=+5.507506910,LastTimestamp:2026-03-18 10:12:46.015772585 +0000 UTC m=+5.507506910,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.515569 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd5c5dadb7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:46.177275319 +0000 UTC m=+5.669009644,LastTimestamp:2026-03-18 10:12:46.177275319 +0000 UTC m=+5.669009644,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.522750 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189de7dd5d2a50a1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:46.190686369 +0000 UTC m=+5.682420694,LastTimestamp:2026-03-18 10:12:46.190686369 +0000 UTC m=+5.682420694,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.535637 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 10:13:05 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d552a46 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 10:13:05 crc kubenswrapper[4733]: body: Mar 18 10:13:05 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.78342919 +0000 UTC m=+14.275163545,LastTimestamp:2026-03-18 10:12:54.78342919 +0000 UTC m=+14.275163545,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:05 crc kubenswrapper[4733]: > Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.541596 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d565238 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.783504952 +0000 UTC m=+14.275239317,LastTimestamp:2026-03-18 10:12:54.783504952 +0000 UTC m=+14.275239317,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.547955 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 10:13:05 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-apiserver-crc.189de7df6ab053e2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 10:13:05 crc kubenswrapper[4733]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 10:13:05 crc kubenswrapper[4733]: Mar 18 10:13:05 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:55.007507426 +0000 UTC m=+14.499241801,LastTimestamp:2026-03-18 10:12:55.007507426 +0000 UTC m=+14.499241801,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:05 crc kubenswrapper[4733]: > Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.554496 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7df6ab3c578 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:55.007733112 +0000 UTC m=+14.499467477,LastTimestamp:2026-03-18 10:12:55.007733112 +0000 UTC m=+14.499467477,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.561829 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189de7df6ab053e2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 10:13:05 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-apiserver-crc.189de7df6ab053e2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 10:13:05 crc kubenswrapper[4733]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 10:13:05 crc kubenswrapper[4733]: Mar 18 10:13:05 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:55.007507426 +0000 UTC m=+14.499241801,LastTimestamp:2026-03-18 10:12:55.01878875 +0000 UTC m=+14.510523115,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:05 crc kubenswrapper[4733]: > Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.570237 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189de7df6ab3c578\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7df6ab3c578 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:55.007733112 +0000 UTC m=+14.499467477,LastTimestamp:2026-03-18 10:12:55.018986246 +0000 UTC m=+14.510720611,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.577508 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189de7dce0dd72da\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dce0dd72da openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.105274074 +0000 UTC m=+3.597008399,LastTimestamp:2026-03-18 10:12:55.279168555 +0000 UTC m=+14.770902880,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.583441 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189de7dced2cd1a0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dced2cd1a0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.311802272 +0000 UTC m=+3.803536597,LastTimestamp:2026-03-18 10:12:55.491477155 +0000 UTC m=+14.983211520,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.588772 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189de7dcee58714f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189de7dcee58714f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:44.331438415 +0000 UTC m=+3.823172740,LastTimestamp:2026-03-18 10:12:55.501787814 +0000 UTC m=+14.993522149,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.596531 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7df5d552a46\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 10:13:05 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d552a46 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 10:13:05 crc kubenswrapper[4733]: body: Mar 18 10:13:05 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.78342919 +0000 UTC m=+14.275163545,LastTimestamp:2026-03-18 10:13:04.78306823 +0000 UTC m=+24.274802555,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:05 crc kubenswrapper[4733]: > Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.601485 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7df5d565238\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d565238 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.783504952 +0000 UTC m=+14.275239317,LastTimestamp:2026-03-18 10:13:04.783148033 +0000 UTC m=+24.274882358,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:05 crc kubenswrapper[4733]: W0318 10:13:05.964158 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 18 10:13:05 crc kubenswrapper[4733]: E0318 10:13:05.964279 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:06 crc kubenswrapper[4733]: I0318 10:13:06.125512 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:07 crc kubenswrapper[4733]: I0318 10:13:07.128272 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.124119 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.407930 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.410781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.410848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.410868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:08 crc kubenswrapper[4733]: I0318 10:13:08.410906 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:08 crc kubenswrapper[4733]: E0318 10:13:08.416615 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:08 crc kubenswrapper[4733]: E0318 10:13:08.416662 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:09 crc kubenswrapper[4733]: I0318 10:13:09.128263 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:09 crc kubenswrapper[4733]: W0318 10:13:09.615579 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 18 10:13:09 crc kubenswrapper[4733]: E0318 10:13:09.615680 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:10 crc kubenswrapper[4733]: I0318 10:13:10.126737 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:10 crc kubenswrapper[4733]: W0318 10:13:10.577849 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:10 crc kubenswrapper[4733]: E0318 10:13:10.577950 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:11 crc kubenswrapper[4733]: I0318 10:13:11.124164 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:11 crc kubenswrapper[4733]: E0318 10:13:11.247593 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:12 crc kubenswrapper[4733]: I0318 10:13:12.125443 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.125879 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.176150 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.177935 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.177998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.178018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.178877 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.502989 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37322->192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.503097 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37322->192.168.126.11:10357: read: connection reset by peer" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.503239 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.503528 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.505626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.505715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.505737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.506686 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 18 10:13:13 crc kubenswrapper[4733]: I0318 10:13:13.506988 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff" gracePeriod=30 Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.517285 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 10:13:13 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-controller-manager-crc.189de7e3b91c203a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:37322->192.168.126.11:10357: read: connection reset by peer Mar 18 10:13:13 crc kubenswrapper[4733]: body: Mar 18 10:13:13 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:13:13.503064122 +0000 UTC m=+32.994798497,LastTimestamp:2026-03-18 10:13:13.503064122 +0000 UTC m=+32.994798497,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:13 crc kubenswrapper[4733]: > Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.523594 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7e3b91d6b0e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37322->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:13:13.503148814 +0000 UTC m=+32.994883179,LastTimestamp:2026-03-18 10:13:13.503148814 +0000 UTC m=+32.994883179,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.534726 4733 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7e3b957a10c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:13:13.506963724 +0000 UTC m=+32.998698089,LastTimestamp:2026-03-18 10:13:13.506963724 +0000 UTC m=+32.998698089,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.542095 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7dc70bd9cce\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc70bd9cce openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.22413947 +0000 UTC m=+1.715873795,LastTimestamp:2026-03-18 10:13:13.526273978 +0000 UTC m=+33.018008353,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.737968 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7dc8408c6ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc8408c6ec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.547832556 +0000 UTC m=+2.039566881,LastTimestamp:2026-03-18 10:13:13.729688945 +0000 UTC m=+33.221423270,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:13 crc kubenswrapper[4733]: E0318 10:13:13.749813 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7dc84ee1035\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7dc84ee1035 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:42.562859061 +0000 UTC m=+2.054593386,LastTimestamp:2026-03-18 10:13:13.739650195 +0000 UTC m=+33.231384520,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.123504 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.352418 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.355732 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11"} Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.355956 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.357457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.357515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.357532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.361704 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.362394 4733 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff" exitCode=255 Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.362465 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff"} Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.362518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2"} Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.362671 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.363852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.363895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:14 crc kubenswrapper[4733]: I0318 10:13:14.363913 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.125940 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.368562 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.369534 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.372601 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" exitCode=255 Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.372668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11"} Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.372718 4733 scope.go:117] "RemoveContainer" containerID="512d400fdc468389180501b48b185ef2c56dbc18c94fa4a8dbd0c2ea829f8c95" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.373006 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.374445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.374508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.374531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.375556 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:15 crc kubenswrapper[4733]: E0318 10:13:15.375849 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.417547 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.419395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.419451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.419469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:15 crc kubenswrapper[4733]: I0318 10:13:15.419507 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:15 crc kubenswrapper[4733]: E0318 10:13:15.423382 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:15 crc kubenswrapper[4733]: E0318 10:13:15.423735 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.124047 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.378525 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.515056 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.515292 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.516613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.516692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.516717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:16 crc kubenswrapper[4733]: I0318 10:13:16.517757 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:16 crc kubenswrapper[4733]: E0318 10:13:16.518076 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:17 crc kubenswrapper[4733]: I0318 10:13:17.127306 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.124980 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.316358 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.316642 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.318357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.318409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.318423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.319073 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:18 crc kubenswrapper[4733]: E0318 10:13:18.319318 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.708813 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.709422 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.711176 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.711280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:18 crc kubenswrapper[4733]: I0318 10:13:18.711299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:19 crc kubenswrapper[4733]: I0318 10:13:19.125090 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:20 crc kubenswrapper[4733]: I0318 10:13:20.125825 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.125775 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:21 crc kubenswrapper[4733]: E0318 10:13:21.247847 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.782377 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.782599 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.784612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.784663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:21 crc kubenswrapper[4733]: I0318 10:13:21.784698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.125415 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.424109 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.426018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.426092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.426110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:22 crc kubenswrapper[4733]: I0318 10:13:22.426157 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:22 crc kubenswrapper[4733]: E0318 10:13:22.431109 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:22 crc kubenswrapper[4733]: E0318 10:13:22.431351 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:23 crc kubenswrapper[4733]: I0318 10:13:23.125281 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:24 crc kubenswrapper[4733]: I0318 10:13:24.123417 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:24 crc kubenswrapper[4733]: I0318 10:13:24.782629 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:13:24 crc kubenswrapper[4733]: I0318 10:13:24.782735 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:13:24 crc kubenswrapper[4733]: E0318 10:13:24.790344 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7df5d552a46\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 10:13:24 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d552a46 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 10:13:24 crc kubenswrapper[4733]: body: Mar 18 10:13:24 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.78342919 +0000 UTC m=+14.275163545,LastTimestamp:2026-03-18 10:13:24.782698705 +0000 UTC m=+44.274433070,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:24 crc kubenswrapper[4733]: > Mar 18 10:13:24 crc kubenswrapper[4733]: E0318 10:13:24.796850 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7df5d565238\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d565238 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.783504952 +0000 UTC m=+14.275239317,LastTimestamp:2026-03-18 10:13:24.782769797 +0000 UTC m=+44.274504162,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:13:25 crc kubenswrapper[4733]: I0318 10:13:25.123372 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:26 crc kubenswrapper[4733]: I0318 10:13:26.124273 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:27 crc kubenswrapper[4733]: I0318 10:13:27.124245 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:27 crc kubenswrapper[4733]: W0318 10:13:27.474138 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 18 10:13:27 crc kubenswrapper[4733]: E0318 10:13:27.474262 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:28 crc kubenswrapper[4733]: I0318 10:13:28.126611 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.125957 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.432362 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.434507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.434591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.434616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:29 crc kubenswrapper[4733]: I0318 10:13:29.434672 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:29 crc kubenswrapper[4733]: E0318 10:13:29.440720 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:29 crc kubenswrapper[4733]: E0318 10:13:29.440804 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:29 crc kubenswrapper[4733]: W0318 10:13:29.890441 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 18 10:13:29 crc kubenswrapper[4733]: E0318 10:13:29.890530 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.126391 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.175180 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.177235 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.177324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.177349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:30 crc kubenswrapper[4733]: I0318 10:13:30.178742 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:30 crc kubenswrapper[4733]: E0318 10:13:30.179118 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:31 crc kubenswrapper[4733]: I0318 10:13:31.124860 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:31 crc kubenswrapper[4733]: E0318 10:13:31.248636 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:31 crc kubenswrapper[4733]: W0318 10:13:31.693606 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 18 10:13:31 crc kubenswrapper[4733]: E0318 10:13:31.693710 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:32 crc kubenswrapper[4733]: W0318 10:13:32.080474 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:32 crc kubenswrapper[4733]: E0318 10:13:32.080581 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.127510 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.861608 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.861918 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.863949 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.864305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:32 crc kubenswrapper[4733]: I0318 10:13:32.864328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:33 crc kubenswrapper[4733]: I0318 10:13:33.128128 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:34 crc kubenswrapper[4733]: I0318 10:13:34.125935 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:34 crc kubenswrapper[4733]: I0318 10:13:34.782968 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:13:34 crc kubenswrapper[4733]: I0318 10:13:34.783063 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:13:34 crc kubenswrapper[4733]: E0318 10:13:34.787765 4733 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189de7df5d552a46\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 10:13:34 crc kubenswrapper[4733]: &Event{ObjectMeta:{kube-controller-manager-crc.189de7df5d552a46 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 10:13:34 crc kubenswrapper[4733]: body: Mar 18 10:13:34 crc kubenswrapper[4733]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:12:54.78342919 +0000 UTC m=+14.275163545,LastTimestamp:2026-03-18 10:13:34.783039193 +0000 UTC m=+54.274773558,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 10:13:34 crc kubenswrapper[4733]: > Mar 18 10:13:35 crc kubenswrapper[4733]: I0318 10:13:35.126251 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.126016 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.440856 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.441714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.441766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.441783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:36 crc kubenswrapper[4733]: I0318 10:13:36.441817 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:36 crc kubenswrapper[4733]: E0318 10:13:36.447156 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:36 crc kubenswrapper[4733]: E0318 10:13:36.447347 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:37 crc kubenswrapper[4733]: I0318 10:13:37.128396 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:38 crc kubenswrapper[4733]: I0318 10:13:38.125158 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:39 crc kubenswrapper[4733]: I0318 10:13:39.123246 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:40 crc kubenswrapper[4733]: I0318 10:13:40.122819 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.123350 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.174920 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.176718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.176759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.176783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.181995 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:41 crc kubenswrapper[4733]: E0318 10:13:41.249302 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.456976 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.458970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38"} Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.459148 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.460627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.460670 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.460681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.786445 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.786641 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.787773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.787812 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.787825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:41 crc kubenswrapper[4733]: I0318 10:13:41.789842 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.122682 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.463000 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.463998 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.465589 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" exitCode=255 Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.465658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38"} Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.465771 4733 scope.go:117] "RemoveContainer" containerID="c5df0e453e549f1c53a257294fcfd5535a89c6524d17c1ea699e8a9a21a19a11" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.465945 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.465954 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467285 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467454 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:42 crc kubenswrapper[4733]: I0318 10:13:42.467837 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:13:42 crc kubenswrapper[4733]: E0318 10:13:42.467993 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.124278 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.447864 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.449287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.449349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.449366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.449401 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:43 crc kubenswrapper[4733]: E0318 10:13:43.453744 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 10:13:43 crc kubenswrapper[4733]: E0318 10:13:43.454417 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 10:13:43 crc kubenswrapper[4733]: I0318 10:13:43.470038 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 10:13:44 crc kubenswrapper[4733]: I0318 10:13:44.123734 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:45 crc kubenswrapper[4733]: I0318 10:13:45.122968 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 10:13:45 crc kubenswrapper[4733]: I0318 10:13:45.575438 4733 csr.go:261] certificate signing request csr-4ptw7 is approved, waiting to be issued Mar 18 10:13:45 crc kubenswrapper[4733]: I0318 10:13:45.582410 4733 csr.go:257] certificate signing request csr-4ptw7 is issued Mar 18 10:13:45 crc kubenswrapper[4733]: I0318 10:13:45.643786 4733 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 18 10:13:45 crc kubenswrapper[4733]: I0318 10:13:45.979242 4733 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.515054 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.515228 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.516397 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.516427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.516439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.517033 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:13:46 crc kubenswrapper[4733]: E0318 10:13:46.517220 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.583378 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-19 14:06:01.863061398 +0000 UTC Mar 18 10:13:46 crc kubenswrapper[4733]: I0318 10:13:46.583448 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 5907h52m15.279618985s for next certificate rotation Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.316783 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.316952 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.318239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.318284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.318293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:48 crc kubenswrapper[4733]: I0318 10:13:48.318875 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:13:48 crc kubenswrapper[4733]: E0318 10:13:48.319065 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.454292 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.455985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.456044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.456070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.456224 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.465881 4733 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.466305 4733 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.466346 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.469823 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.469872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.469886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.469904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.469919 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:13:50Z","lastTransitionTime":"2026-03-18T10:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.487077 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.498730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.498810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.498830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.498850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.498864 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:13:50Z","lastTransitionTime":"2026-03-18T10:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.526705 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.538745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.538777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.538784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.538799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.538809 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:13:50Z","lastTransitionTime":"2026-03-18T10:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.552599 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.561148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.561181 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.561205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.561223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:13:50 crc kubenswrapper[4733]: I0318 10:13:50.561232 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:13:50Z","lastTransitionTime":"2026-03-18T10:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.574674 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.574799 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.574825 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.675427 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.776468 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.877538 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:50 crc kubenswrapper[4733]: E0318 10:13:50.978145 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.078278 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.179308 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.249862 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.280235 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.381041 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.482528 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.583556 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.684408 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.785426 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.887002 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:51 crc kubenswrapper[4733]: E0318 10:13:51.988299 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.089507 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.190336 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.290764 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.391875 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.492023 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.595382 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.696748 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.797117 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.898059 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:52 crc kubenswrapper[4733]: E0318 10:13:52.998899 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.099562 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.200145 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.300435 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.401165 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.502311 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.602503 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.703247 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.803562 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:53 crc kubenswrapper[4733]: E0318 10:13:53.904631 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.005354 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.105617 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.206520 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.307723 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.408699 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.508843 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.609750 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.710877 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.811616 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:54 crc kubenswrapper[4733]: E0318 10:13:54.912738 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.013786 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.114930 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.215765 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.316237 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.417364 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.518261 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.619138 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.720612 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.821349 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:55 crc kubenswrapper[4733]: E0318 10:13:55.922263 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.022584 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.123673 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.225030 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.326072 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.426559 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.527537 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.627936 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.728759 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.829710 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:56 crc kubenswrapper[4733]: E0318 10:13:56.930605 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.031670 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.132077 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.232470 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.333591 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.433664 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.534107 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.634239 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.734403 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.835271 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:57 crc kubenswrapper[4733]: E0318 10:13:57.936226 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.036336 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.136842 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.237436 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.337787 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.437882 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.538461 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.639516 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: I0318 10:13:58.694941 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.740643 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.841836 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:58 crc kubenswrapper[4733]: E0318 10:13:58.942863 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.043650 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.144789 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.245681 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.346407 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.446580 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.547392 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.647874 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.748831 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.849343 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:13:59 crc kubenswrapper[4733]: E0318 10:13:59.950101 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.051157 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.152155 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.175395 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.176513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.176598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.176622 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.177152 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.177328 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.253152 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.353455 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.454155 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.554869 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.655739 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.756810 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.838601 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.843746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.843786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.843802 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.843821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.843834 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:00Z","lastTransitionTime":"2026-03-18T10:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.858652 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.863239 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.863269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.863280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.863299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.863311 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:00Z","lastTransitionTime":"2026-03-18T10:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.877967 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.883599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.883662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.883675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.883694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.883710 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:00Z","lastTransitionTime":"2026-03-18T10:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.899025 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.903133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.903175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.903209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.903227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:00 crc kubenswrapper[4733]: I0318 10:14:00.903240 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:00Z","lastTransitionTime":"2026-03-18T10:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.917851 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.918001 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:00 crc kubenswrapper[4733]: E0318 10:14:00.918037 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.018616 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.119145 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.219707 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.250373 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.320589 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.421453 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.522445 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.622715 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.723023 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.823377 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:01 crc kubenswrapper[4733]: E0318 10:14:01.923704 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.024163 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.124716 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.225776 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.326260 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.426855 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.527015 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.627700 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.729042 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.830019 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:02 crc kubenswrapper[4733]: E0318 10:14:02.930923 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.031137 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.132106 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.233280 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.333921 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.435123 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.536062 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.637337 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.738622 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.839477 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:03 crc kubenswrapper[4733]: E0318 10:14:03.940121 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.040261 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.141316 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.241859 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.342167 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.442732 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.543464 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.644282 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.762865 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.863910 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:04 crc kubenswrapper[4733]: E0318 10:14:04.964899 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.065895 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.167167 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.268280 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.369384 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.470540 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.570885 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.672059 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.772801 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.873965 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:05 crc kubenswrapper[4733]: E0318 10:14:05.974817 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.075814 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.177060 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.278274 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.379030 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.479526 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.580090 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.681199 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.782102 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.883040 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:06 crc kubenswrapper[4733]: E0318 10:14:06.984082 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.085264 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.185670 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.286501 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.387511 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.487985 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.589121 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.689710 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.790339 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.891497 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:07 crc kubenswrapper[4733]: E0318 10:14:07.992262 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.093347 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.193981 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.294963 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.395479 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.496543 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.597585 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.698291 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.798593 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.899765 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:08 crc kubenswrapper[4733]: E0318 10:14:08.999906 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.100088 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.200649 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.301459 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.402648 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.503008 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.604122 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.704296 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.804454 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:09 crc kubenswrapper[4733]: E0318 10:14:09.905532 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.006136 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.106561 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.207395 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.308219 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.409263 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.509422 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.609958 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.710846 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.811133 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:10 crc kubenswrapper[4733]: E0318 10:14:10.912049 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.012564 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.113571 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.116904 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.121553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.121596 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.121609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.121628 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.121642 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:11Z","lastTransitionTime":"2026-03-18T10:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.130951 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.134659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.134677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.134685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.134698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.134706 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:11Z","lastTransitionTime":"2026-03-18T10:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.143347 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.146346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.146361 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.146370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.146382 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.146391 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:11Z","lastTransitionTime":"2026-03-18T10:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.154683 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.158933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.158966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.158977 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.158995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.159007 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:11Z","lastTransitionTime":"2026-03-18T10:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.169658 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.169809 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.174634 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.175524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.175579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.175591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:11 crc kubenswrapper[4733]: I0318 10:14:11.176290 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.176522 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.214013 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.250710 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.314944 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.416008 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.516593 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.617035 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.718061 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.818702 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:11 crc kubenswrapper[4733]: E0318 10:14:11.919612 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.020731 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.121083 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.221526 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.321639 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.422732 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.523377 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.623563 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.723961 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.824991 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:12 crc kubenswrapper[4733]: E0318 10:14:12.925529 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.026697 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.048336 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.129383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.129446 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.129469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.129497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.129522 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.158685 4733 apiserver.go:52] "Watching apiserver" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.167173 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.168796 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-7pxwd","openshift-multus/multus-additional-cni-plugins-t28sh","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-image-registry/node-ca-xfvfl","openshift-multus/multus-g6j2q","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-hsk58","openshift-multus/network-metrics-daemon-4s425","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-machine-config-operator/machine-config-daemon-2h7dp"] Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.169280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.169289 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.169486 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.169900 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.172111 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.172136 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.170765 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.172221 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.170982 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.172303 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.171060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.170877 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.172705 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.169906 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.171019 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.171171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.170149 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.171321 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.173481 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.174221 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.174986 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.175983 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.176021 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.177622 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.181903 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.182267 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.182495 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.184860 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.184950 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.184870 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.184962 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.185540 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.185649 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.185740 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.185966 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.185979 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.186109 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.186239 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.186256 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.186553 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.186558 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.188021 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.188945 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.189416 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.189732 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.190013 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.190287 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.190612 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.190850 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.191067 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.191928 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.192244 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.192543 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.192867 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.194531 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.200689 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.204253 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.222227 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-hosts-file\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.222476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.222514 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9zpb\" (UniqueName: \"kubernetes.io/projected/b3650177-e338-4eba-ab42-bc0cd14c9d65-kube-api-access-x9zpb\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.222543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-httph\" (UniqueName: \"kubernetes.io/projected/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-kube-api-access-httph\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.223522 4733 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.223759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.233295 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.233336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.233354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.233379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.233399 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.241174 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.253665 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.266072 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.280452 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.288819 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.297732 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.311856 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323104 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323166 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323286 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323325 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323363 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323398 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323463 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323495 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323530 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323560 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323627 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323723 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323758 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323793 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323912 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.323963 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324001 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324036 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324082 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324119 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324152 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324208 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324260 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324426 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324468 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324504 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324541 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324656 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324722 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324772 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324806 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324844 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324881 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324913 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324947 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.324981 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325057 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325130 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325283 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325528 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325457 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325652 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325687 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325721 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325758 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325882 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325957 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326029 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326066 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326099 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326131 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326164 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326221 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326255 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326285 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326315 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326346 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326377 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326476 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326542 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326580 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326612 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326643 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326677 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326743 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326777 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326810 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326758 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327010 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325884 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.325969 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326144 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326379 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326415 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326959 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.326949 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327286 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327600 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327625 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327657 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327755 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327783 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327796 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.327831 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328171 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328168 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328119 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328124 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328384 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328494 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328734 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328780 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.328960 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329120 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329578 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329618 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.329737 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.828074935 +0000 UTC m=+93.319809340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330776 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329757 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329833 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329871 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329971 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.329978 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330154 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330153 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330373 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330514 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330808 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.330826 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331328 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331335 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331444 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331464 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331686 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331729 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.331721 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332214 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332261 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332357 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332449 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332658 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.332726 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333033 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333110 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333392 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333433 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333458 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333518 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333517 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333563 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333808 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.333872 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334023 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334303 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334314 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334524 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334451 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334607 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334624 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334748 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.334995 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335217 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335004 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335518 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335592 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335666 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335708 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335723 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335768 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335805 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335840 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335908 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.335985 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336039 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336369 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336294 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336673 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336691 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336700 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.336755 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337050 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337128 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337463 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337603 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337645 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337662 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.337863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338331 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338423 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338506 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338760 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338825 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338759 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338858 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338928 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.338970 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339009 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339048 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339214 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339256 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339295 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339328 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339331 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339356 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339466 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339508 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339543 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339582 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339615 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339734 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339777 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339818 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339828 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339819 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340023 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340054 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340089 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340123 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340238 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340271 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340305 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340344 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340376 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340407 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340440 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340473 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340536 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340605 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340670 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340700 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340733 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340764 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340800 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340831 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340865 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340899 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340933 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340966 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341002 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341035 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341070 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341104 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341137 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341171 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341241 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341277 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341311 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341343 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341377 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341414 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341448 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341482 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.339945 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340068 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340209 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340405 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340443 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340481 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.340758 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341018 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341098 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341311 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341552 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341580 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341876 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.341984 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342009 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342021 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342029 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342161 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342336 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342349 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342369 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342397 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342444 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342494 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342517 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342540 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342564 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342586 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342607 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342633 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342662 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342698 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342725 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342748 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342781 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342825 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342856 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342892 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342922 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342947 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342949 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342964 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343009 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343326 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwk4s\" (UniqueName: \"kubernetes.io/projected/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-kube-api-access-xwk4s\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343382 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-httph\" (UniqueName: \"kubernetes.io/projected/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-kube-api-access-httph\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9zpb\" (UniqueName: \"kubernetes.io/projected/b3650177-e338-4eba-ab42-bc0cd14c9d65-kube-api-access-x9zpb\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343444 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343460 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bb58b528-9013-4fab-9747-60bb6ff1bc1f-serviceca\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343504 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cnibin\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343545 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-kubelet\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343579 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-conf-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343610 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb58b528-9013-4fab-9747-60bb6ff1bc1f-host\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343674 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f75e1c5-e0c5-43df-944f-77b734070793-rootfs\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343703 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f75e1c5-e0c5-43df-944f-77b734070793-proxy-tls\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343737 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-binary-copy\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343771 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343805 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg7jp\" (UniqueName: \"kubernetes.io/projected/bb58b528-9013-4fab-9747-60bb6ff1bc1f-kube-api-access-zg7jp\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343837 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cnibin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343870 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-os-release\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343902 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-multus-certs\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343933 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343966 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344001 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqxdr\" (UniqueName: \"kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344035 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344072 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/6f75e1c5-e0c5-43df-944f-77b734070793-kube-api-access-xpnv6\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-os-release\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344138 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-hosts-file\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344234 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344271 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344305 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344336 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-k8s-cni-cncf-io\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344373 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-multus\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344412 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-socket-dir-parent\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344449 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d693a73-68c1-4595-bbcc-be97691b06fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344512 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-system-cni-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344612 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cni-binary-copy\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344643 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-bin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344675 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-etc-kubernetes\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344706 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph8vv\" (UniqueName: \"kubernetes.io/projected/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-kube-api-access-ph8vv\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344746 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344779 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344810 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.342328 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343509 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343713 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.343877 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344182 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344336 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344370 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344739 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344889 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345387 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345403 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345858 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345932 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-hosts-file\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.345957 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.346235 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.346285 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.346375 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.346519 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.346613 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.846588872 +0000 UTC m=+93.338323227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.346950 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.347472 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.347741 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.349183 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.349949 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.350312 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.350671 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.351788 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.351802 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.351853 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.360350 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.362082 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.362329 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.362483 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.362966 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.363081 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.363588 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.363825 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.363921 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.364352 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.364360 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.364965 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.365043 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.365064 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.365453 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.365778 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.365822 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366180 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366391 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.344896 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-system-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366767 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366803 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366814 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366824 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366879 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366920 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367026 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367074 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367090 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367178 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f75e1c5-e0c5-43df-944f-77b734070793-mcd-auth-proxy-config\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367225 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367275 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367273 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367307 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg7hc\" (UniqueName: \"kubernetes.io/projected/7d693a73-68c1-4595-bbcc-be97691b06fe-kube-api-access-vg7hc\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367450 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367504 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367572 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367720 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367737 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367796 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367898 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.367999 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-netns\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368068 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-hostroot\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368123 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-daemon-config\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368176 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.366580 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368271 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368328 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368379 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368782 4733 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368833 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368841 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368867 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368898 4733 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368929 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368959 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.368987 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369016 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369046 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369078 4733 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369107 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369149 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369185 4733 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.369235 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.369317 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.869295816 +0000 UTC m=+93.361030161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369242 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369631 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369648 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369664 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369679 4733 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369692 4733 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369707 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369720 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369735 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369748 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369762 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369775 4733 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369788 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369800 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369813 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369826 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369838 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.370008 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.370050 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.870038077 +0000 UTC m=+93.361772412 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370168 4733 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370481 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.369853 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370541 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370562 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370582 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370600 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370624 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370667 4733 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370687 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370707 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370725 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370744 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370762 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370784 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370802 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370820 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370837 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370854 4733 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370873 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370892 4733 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370910 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370927 4733 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370945 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370963 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.370981 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371001 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371018 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371035 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371052 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371066 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371078 4733 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371091 4733 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371103 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371116 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371128 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371131 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371141 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371154 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371168 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371181 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371224 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371242 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371261 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371276 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371288 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371301 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371315 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371327 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371340 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371353 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371365 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371378 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371390 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371403 4733 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371415 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371429 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371441 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371454 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371468 4733 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371479 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371492 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371505 4733 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371524 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371535 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371549 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371562 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371574 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371587 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371603 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371616 4733 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371628 4733 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371641 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371652 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371664 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371676 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371688 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371701 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371713 4733 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371728 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371740 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371752 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371765 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371777 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371789 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371802 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371813 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371826 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371839 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371851 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371863 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371876 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371888 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371900 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371913 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371927 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371939 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371950 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371962 4733 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371974 4733 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371985 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.371997 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372008 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372020 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372032 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372044 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372056 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372067 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372078 4733 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372090 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372101 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372113 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372125 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372137 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372150 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372163 4733 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372174 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372206 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372224 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372240 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372257 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372273 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372291 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372306 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372325 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372342 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372354 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372448 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372465 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372486 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372499 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372513 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372599 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372635 4733 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372658 4733 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372679 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372700 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372724 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.372731 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.372786 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.372815 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.372963 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.872921257 +0000 UTC m=+93.364655792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.372743 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373041 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373067 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373090 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373111 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373139 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.373163 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.383110 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.383554 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.383769 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.385436 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.385162 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.387949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.388357 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.388872 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.391482 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.391609 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.392714 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.392748 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.392772 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.392871 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:13.892842293 +0000 UTC m=+93.384576818 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.392990 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9zpb\" (UniqueName: \"kubernetes.io/projected/b3650177-e338-4eba-ab42-bc0cd14c9d65-kube-api-access-x9zpb\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.393426 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.393969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-httph\" (UniqueName: \"kubernetes.io/projected/c2c181c8-3361-40a2-afc5-a677e0ab4ecd-kube-api-access-httph\") pod \"node-resolver-hsk58\" (UID: \"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\") " pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.396335 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.397706 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.398521 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.399043 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.399058 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.399258 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.399386 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.399155 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.400357 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.400747 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.404636 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.407146 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.417363 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.420870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.427537 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.432810 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.433049 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.439790 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.442551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.442599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.442611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.442632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.442647 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.447556 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.456675 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474166 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474205 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474242 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f75e1c5-e0c5-43df-944f-77b734070793-mcd-auth-proxy-config\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474263 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg7hc\" (UniqueName: \"kubernetes.io/projected/7d693a73-68c1-4595-bbcc-be97691b06fe-kube-api-access-vg7hc\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474321 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474342 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474495 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474722 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-netns\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.474909 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-netns\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-env-overrides\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475384 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-hostroot\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475362 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-daemon-config\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475491 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475517 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475490 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-hostroot\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475541 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475549 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475565 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d693a73-68c1-4595-bbcc-be97691b06fe-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475678 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475702 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwk4s\" (UniqueName: \"kubernetes.io/projected/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-kube-api-access-xwk4s\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.475299 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f75e1c5-e0c5-43df-944f-77b734070793-mcd-auth-proxy-config\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476017 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bb58b528-9013-4fab-9747-60bb6ff1bc1f-serviceca\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-kubelet\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-conf-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cnibin\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb58b528-9013-4fab-9747-60bb6ff1bc1f-host\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476240 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476274 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f75e1c5-e0c5-43df-944f-77b734070793-proxy-tls\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-binary-copy\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476317 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-conf-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476327 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg7jp\" (UniqueName: \"kubernetes.io/projected/bb58b528-9013-4fab-9747-60bb6ff1bc1f-kube-api-access-zg7jp\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476481 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cnibin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476532 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-os-release\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f75e1c5-e0c5-43df-944f-77b734070793-rootfs\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476245 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-kubelet\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-multus-certs\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476671 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-multus-certs\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476708 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqxdr\" (UniqueName: \"kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cnibin\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476621 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-daemon-config\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476793 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-os-release\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476805 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476676 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb58b528-9013-4fab-9747-60bb6ff1bc1f-host\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f75e1c5-e0c5-43df-944f-77b734070793-rootfs\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476734 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cnibin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476755 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/6f75e1c5-e0c5-43df-944f-77b734070793-kube-api-access-xpnv6\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476912 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-os-release\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476954 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-k8s-cni-cncf-io\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.476858 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477032 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-multus\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477081 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-socket-dir-parent\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477104 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477139 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477159 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d693a73-68c1-4595-bbcc-be97691b06fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477172 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-os-release\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477180 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-system-cni-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477230 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cni-binary-copy\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-bin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477291 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-etc-kubernetes\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477308 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-multus-socket-dir-parent\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477353 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477412 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-multus\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477452 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-etc-kubernetes\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477487 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-system-cni-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477512 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-run-k8s-cni-cncf-io\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477714 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477787 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-system-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph8vv\" (UniqueName: \"kubernetes.io/projected/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-kube-api-access-ph8vv\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-system-cni-dir\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.477983 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-host-var-lib-cni-bin\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.478271 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.478332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.478600 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.478644 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-cni-binary-copy\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479349 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479586 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479597 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479608 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479617 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479627 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479636 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479646 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479655 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479666 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479675 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479685 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479694 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479704 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479715 4733 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479725 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479737 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479748 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479758 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479768 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479780 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.479803 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bb58b528-9013-4fab-9747-60bb6ff1bc1f-serviceca\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.480533 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-cni-binary-copy\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.489916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.490251 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f75e1c5-e0c5-43df-944f-77b734070793-proxy-tls\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.491348 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d693a73-68c1-4595-bbcc-be97691b06fe-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.493105 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqxdr\" (UniqueName: \"kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr\") pod \"ovnkube-node-7pxwd\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.494523 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwk4s\" (UniqueName: \"kubernetes.io/projected/0f82588a-9dbd-4c55-8cfc-f96e57fa58b9-kube-api-access-xwk4s\") pod \"multus-additional-cni-plugins-t28sh\" (UID: \"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\") " pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.495822 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph8vv\" (UniqueName: \"kubernetes.io/projected/cc85b0d4-15a5-4894-9f07-9aaeb28f63fa-kube-api-access-ph8vv\") pod \"multus-g6j2q\" (UID: \"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\") " pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.496071 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg7hc\" (UniqueName: \"kubernetes.io/projected/7d693a73-68c1-4595-bbcc-be97691b06fe-kube-api-access-vg7hc\") pod \"ovnkube-control-plane-749d76644c-spfjj\" (UID: \"7d693a73-68c1-4595-bbcc-be97691b06fe\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.497097 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnv6\" (UniqueName: \"kubernetes.io/projected/6f75e1c5-e0c5-43df-944f-77b734070793-kube-api-access-xpnv6\") pod \"machine-config-daemon-2h7dp\" (UID: \"6f75e1c5-e0c5-43df-944f-77b734070793\") " pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.498865 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg7jp\" (UniqueName: \"kubernetes.io/projected/bb58b528-9013-4fab-9747-60bb6ff1bc1f-kube-api-access-zg7jp\") pod \"node-ca-xfvfl\" (UID: \"bb58b528-9013-4fab-9747-60bb6ff1bc1f\") " pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.520774 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.521433 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xfvfl" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.531082 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.536392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t28sh" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.545015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.547639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.547728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.547786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.547860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.547919 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.547794 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e36bf33aebc25dbcc802b0414aa206e2fb77c5c19a7273e84ddfa5ff8f3ed1da WatchSource:0}: Error finding container e36bf33aebc25dbcc802b0414aa206e2fb77c5c19a7273e84ddfa5ff8f3ed1da: Status 404 returned error can't find the container with id e36bf33aebc25dbcc802b0414aa206e2fb77c5c19a7273e84ddfa5ff8f3ed1da Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.552882 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.554846 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:13 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: source /etc/kubernetes/apiserver-url.env Mar 18 10:14:13 crc kubenswrapper[4733]: else Mar 18 10:14:13 crc kubenswrapper[4733]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 18 10:14:13 crc kubenswrapper[4733]: exit 1 Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.554921 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 18 10:14:13 crc kubenswrapper[4733]: while [ true ]; Mar 18 10:14:13 crc kubenswrapper[4733]: do Mar 18 10:14:13 crc kubenswrapper[4733]: for f in $(ls /tmp/serviceca); do Mar 18 10:14:13 crc kubenswrapper[4733]: echo $f Mar 18 10:14:13 crc kubenswrapper[4733]: ca_file_path="/tmp/serviceca/${f}" Mar 18 10:14:13 crc kubenswrapper[4733]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 18 10:14:13 crc kubenswrapper[4733]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 18 10:14:13 crc kubenswrapper[4733]: if [ -e "${reg_dir_path}" ]; then Mar 18 10:14:13 crc kubenswrapper[4733]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 18 10:14:13 crc kubenswrapper[4733]: else Mar 18 10:14:13 crc kubenswrapper[4733]: mkdir $reg_dir_path Mar 18 10:14:13 crc kubenswrapper[4733]: cp $ca_file_path $reg_dir_path/ca.crt Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: for d in $(ls /etc/docker/certs.d); do Mar 18 10:14:13 crc kubenswrapper[4733]: echo $d Mar 18 10:14:13 crc kubenswrapper[4733]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 18 10:14:13 crc kubenswrapper[4733]: reg_conf_path="/tmp/serviceca/${dp}" Mar 18 10:14:13 crc kubenswrapper[4733]: if [ ! -e "${reg_conf_path}" ]; then Mar 18 10:14:13 crc kubenswrapper[4733]: rm -rf /etc/docker/certs.d/$d Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 60 & wait ${!} Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zg7jp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-xfvfl_openshift-image-registry(bb58b528-9013-4fab-9747-60bb6ff1bc1f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.556626 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xfvfl" podUID="bb58b528-9013-4fab-9747-60bb6ff1bc1f" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.556680 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.557802 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hsk58" Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.561762 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73327417_4d3b_45f1_b3b6_575fdeeaa31a.slice/crio-35bea9a3e63456f3c4522f7b18c54f2df3fc823d29bd3059264ea8e5f121d012 WatchSource:0}: Error finding container 35bea9a3e63456f3c4522f7b18c54f2df3fc823d29bd3059264ea8e5f121d012: Status 404 returned error can't find the container with id 35bea9a3e63456f3c4522f7b18c54f2df3fc823d29bd3059264ea8e5f121d012 Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.563714 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.569968 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"35bea9a3e63456f3c4522f7b18c54f2df3fc823d29bd3059264ea8e5f121d012"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.570333 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.570351 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f82588a_9dbd_4c55_8cfc_f96e57fa58b9.slice/crio-ff1d93cef35a316c51b4c062176258993d12b241a2a73aaf584a089109ebf4c1 WatchSource:0}: Error finding container ff1d93cef35a316c51b4c062176258993d12b241a2a73aaf584a089109ebf4c1: Status 404 returned error can't find the container with id ff1d93cef35a316c51b4c062176258993d12b241a2a73aaf584a089109ebf4c1 Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.571284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xfvfl" event={"ID":"bb58b528-9013-4fab-9747-60bb6ff1bc1f","Type":"ContainerStarted","Data":"da169ce5c75cd05bb49cb01f0a3fda10717fc3999af212b34c0dcccc7dbbab26"} Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.572407 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7fb0376f2eec0926827408777c433a08963622c37f222da5c978e502f9dbcbfd WatchSource:0}: Error finding container 7fb0376f2eec0926827408777c433a08963622c37f222da5c978e502f9dbcbfd: Status 404 returned error can't find the container with id 7fb0376f2eec0926827408777c433a08963622c37f222da5c978e502f9dbcbfd Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.572577 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 18 10:14:13 crc kubenswrapper[4733]: apiVersion: v1 Mar 18 10:14:13 crc kubenswrapper[4733]: clusters: Mar 18 10:14:13 crc kubenswrapper[4733]: - cluster: Mar 18 10:14:13 crc kubenswrapper[4733]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 18 10:14:13 crc kubenswrapper[4733]: server: https://api-int.crc.testing:6443 Mar 18 10:14:13 crc kubenswrapper[4733]: name: default-cluster Mar 18 10:14:13 crc kubenswrapper[4733]: contexts: Mar 18 10:14:13 crc kubenswrapper[4733]: - context: Mar 18 10:14:13 crc kubenswrapper[4733]: cluster: default-cluster Mar 18 10:14:13 crc kubenswrapper[4733]: namespace: default Mar 18 10:14:13 crc kubenswrapper[4733]: user: default-auth Mar 18 10:14:13 crc kubenswrapper[4733]: name: default-context Mar 18 10:14:13 crc kubenswrapper[4733]: current-context: default-context Mar 18 10:14:13 crc kubenswrapper[4733]: kind: Config Mar 18 10:14:13 crc kubenswrapper[4733]: preferences: {} Mar 18 10:14:13 crc kubenswrapper[4733]: users: Mar 18 10:14:13 crc kubenswrapper[4733]: - name: default-auth Mar 18 10:14:13 crc kubenswrapper[4733]: user: Mar 18 10:14:13 crc kubenswrapper[4733]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 18 10:14:13 crc kubenswrapper[4733]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 18 10:14:13 crc kubenswrapper[4733]: EOF Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zqxdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.573609 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e36bf33aebc25dbcc802b0414aa206e2fb77c5c19a7273e84ddfa5ff8f3ed1da"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.577312 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g6j2q" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.578112 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:13 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: source /etc/kubernetes/apiserver-url.env Mar 18 10:14:13 crc kubenswrapper[4733]: else Mar 18 10:14:13 crc kubenswrapper[4733]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 18 10:14:13 crc kubenswrapper[4733]: exit 1 Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.578130 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:13 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 18 10:14:13 crc kubenswrapper[4733]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 18 10:14:13 crc kubenswrapper[4733]: ho_enable="--enable-hybrid-overlay" Mar 18 10:14:13 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 18 10:14:13 crc kubenswrapper[4733]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 18 10:14:13 crc kubenswrapper[4733]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 10:14:13 crc kubenswrapper[4733]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --webhook-host=127.0.0.1 \ Mar 18 10:14:13 crc kubenswrapper[4733]: --webhook-port=9743 \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${ho_enable} \ Mar 18 10:14:13 crc kubenswrapper[4733]: --enable-interconnect \ Mar 18 10:14:13 crc kubenswrapper[4733]: --disable-approver \ Mar 18 10:14:13 crc kubenswrapper[4733]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --wait-for-kubernetes-api=200s \ Mar 18 10:14:13 crc kubenswrapper[4733]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --loglevel="${LOGLEVEL}" Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.579072 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 18 10:14:13 crc kubenswrapper[4733]: while [ true ]; Mar 18 10:14:13 crc kubenswrapper[4733]: do Mar 18 10:14:13 crc kubenswrapper[4733]: for f in $(ls /tmp/serviceca); do Mar 18 10:14:13 crc kubenswrapper[4733]: echo $f Mar 18 10:14:13 crc kubenswrapper[4733]: ca_file_path="/tmp/serviceca/${f}" Mar 18 10:14:13 crc kubenswrapper[4733]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 18 10:14:13 crc kubenswrapper[4733]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 18 10:14:13 crc kubenswrapper[4733]: if [ -e "${reg_dir_path}" ]; then Mar 18 10:14:13 crc kubenswrapper[4733]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 18 10:14:13 crc kubenswrapper[4733]: else Mar 18 10:14:13 crc kubenswrapper[4733]: mkdir $reg_dir_path Mar 18 10:14:13 crc kubenswrapper[4733]: cp $ca_file_path $reg_dir_path/ca.crt Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: for d in $(ls /etc/docker/certs.d); do Mar 18 10:14:13 crc kubenswrapper[4733]: echo $d Mar 18 10:14:13 crc kubenswrapper[4733]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 18 10:14:13 crc kubenswrapper[4733]: reg_conf_path="/tmp/serviceca/${dp}" Mar 18 10:14:13 crc kubenswrapper[4733]: if [ ! -e "${reg_conf_path}" ]; then Mar 18 10:14:13 crc kubenswrapper[4733]: rm -rf /etc/docker/certs.d/$d Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 60 & wait ${!} Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zg7jp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-xfvfl_openshift-image-registry(bb58b528-9013-4fab-9747-60bb6ff1bc1f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.579337 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.580368 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xfvfl" podUID="bb58b528-9013-4fab-9747-60bb6ff1bc1f" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.580410 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.580659 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.583244 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwk4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-t28sh_openshift-multus(0f82588a-9dbd-4c55-8cfc-f96e57fa58b9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.584564 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-t28sh" podUID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.585529 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:13 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 10:14:13 crc kubenswrapper[4733]: --disable-webhook \ Mar 18 10:14:13 crc kubenswrapper[4733]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --loglevel="${LOGLEVEL}" Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.586842 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.590580 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.604095 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.604975 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-082c29e3548c80a38ab2a902a222a7bd38f378ff323b000dc303bf38651a6b0f WatchSource:0}: Error finding container 082c29e3548c80a38ab2a902a222a7bd38f378ff323b000dc303bf38651a6b0f: Status 404 returned error can't find the container with id 082c29e3548c80a38ab2a902a222a7bd38f378ff323b000dc303bf38651a6b0f Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.606321 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:13 crc kubenswrapper[4733]: set -uo pipefail Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 18 10:14:13 crc kubenswrapper[4733]: HOSTS_FILE="/etc/hosts" Mar 18 10:14:13 crc kubenswrapper[4733]: TEMP_FILE="/etc/hosts.tmp" Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: # Make a temporary file with the old hosts file's attributes. Mar 18 10:14:13 crc kubenswrapper[4733]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 18 10:14:13 crc kubenswrapper[4733]: echo "Failed to preserve hosts file. Exiting." Mar 18 10:14:13 crc kubenswrapper[4733]: exit 1 Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: while true; do Mar 18 10:14:13 crc kubenswrapper[4733]: declare -A svc_ips Mar 18 10:14:13 crc kubenswrapper[4733]: for svc in "${services[@]}"; do Mar 18 10:14:13 crc kubenswrapper[4733]: # Fetch service IP from cluster dns if present. We make several tries Mar 18 10:14:13 crc kubenswrapper[4733]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 18 10:14:13 crc kubenswrapper[4733]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 18 10:14:13 crc kubenswrapper[4733]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 18 10:14:13 crc kubenswrapper[4733]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:13 crc kubenswrapper[4733]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:13 crc kubenswrapper[4733]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:13 crc kubenswrapper[4733]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 18 10:14:13 crc kubenswrapper[4733]: for i in ${!cmds[*]} Mar 18 10:14:13 crc kubenswrapper[4733]: do Mar 18 10:14:13 crc kubenswrapper[4733]: ips=($(eval "${cmds[i]}")) Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: svc_ips["${svc}"]="${ips[@]}" Mar 18 10:14:13 crc kubenswrapper[4733]: break Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: # Update /etc/hosts only if we get valid service IPs Mar 18 10:14:13 crc kubenswrapper[4733]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 18 10:14:13 crc kubenswrapper[4733]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 18 10:14:13 crc kubenswrapper[4733]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 18 10:14:13 crc kubenswrapper[4733]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:13 crc kubenswrapper[4733]: continue Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: # Append resolver entries for services Mar 18 10:14:13 crc kubenswrapper[4733]: rc=0 Mar 18 10:14:13 crc kubenswrapper[4733]: for svc in "${!svc_ips[@]}"; do Mar 18 10:14:13 crc kubenswrapper[4733]: for ip in ${svc_ips[${svc}]}; do Mar 18 10:14:13 crc kubenswrapper[4733]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ $rc -ne 0 ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:13 crc kubenswrapper[4733]: continue Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 18 10:14:13 crc kubenswrapper[4733]: # Replace /etc/hosts with our modified version if needed Mar 18 10:14:13 crc kubenswrapper[4733]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 18 10:14:13 crc kubenswrapper[4733]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:13 crc kubenswrapper[4733]: unset svc_ips Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-httph,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-hsk58_openshift-dns(c2c181c8-3361-40a2-afc5-a677e0ab4ecd): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.606321 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:13 crc kubenswrapper[4733]: set -euo pipefail Mar 18 10:14:13 crc kubenswrapper[4733]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 18 10:14:13 crc kubenswrapper[4733]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 18 10:14:13 crc kubenswrapper[4733]: # As the secret mount is optional we must wait for the files to be present. Mar 18 10:14:13 crc kubenswrapper[4733]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 18 10:14:13 crc kubenswrapper[4733]: TS=$(date +%s) Mar 18 10:14:13 crc kubenswrapper[4733]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 18 10:14:13 crc kubenswrapper[4733]: HAS_LOGGED_INFO=0 Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: log_missing_certs(){ Mar 18 10:14:13 crc kubenswrapper[4733]: CUR_TS=$(date +%s) Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 18 10:14:13 crc kubenswrapper[4733]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 18 10:14:13 crc kubenswrapper[4733]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 18 10:14:13 crc kubenswrapper[4733]: HAS_LOGGED_INFO=1 Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: } Mar 18 10:14:13 crc kubenswrapper[4733]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 18 10:14:13 crc kubenswrapper[4733]: log_missing_certs Mar 18 10:14:13 crc kubenswrapper[4733]: sleep 5 Mar 18 10:14:13 crc kubenswrapper[4733]: done Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/kube-rbac-proxy \ Mar 18 10:14:13 crc kubenswrapper[4733]: --logtostderr \ Mar 18 10:14:13 crc kubenswrapper[4733]: --secure-listen-address=:9108 \ Mar 18 10:14:13 crc kubenswrapper[4733]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 18 10:14:13 crc kubenswrapper[4733]: --upstream=http://127.0.0.1:29108/ \ Mar 18 10:14:13 crc kubenswrapper[4733]: --tls-private-key-file=${TLS_PK} \ Mar 18 10:14:13 crc kubenswrapper[4733]: --tls-cert-file=${TLS_CERT} Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vg7hc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-spfjj_openshift-ovn-kubernetes(7d693a73-68c1-4595-bbcc-be97691b06fe): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.607622 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-hsk58" podUID="c2c181c8-3361-40a2-afc5-a677e0ab4ecd" Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.614360 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f75e1c5_e0c5_43df_944f_77b734070793.slice/crio-44de9df63ec3ee4eae109c26660b6c1eb3f01d1f6242fd4f233e7ef6cfded6a1 WatchSource:0}: Error finding container 44de9df63ec3ee4eae109c26660b6c1eb3f01d1f6242fd4f233e7ef6cfded6a1: Status 404 returned error can't find the container with id 44de9df63ec3ee4eae109c26660b6c1eb3f01d1f6242fd4f233e7ef6cfded6a1 Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.614915 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:13 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v4_join_subnet_opt= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v6_join_subnet_opt= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v4_transit_switch_subnet_opt= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v6_transit_switch_subnet_opt= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: dns_name_resolver_enabled_flag= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "false" == "true" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: persistent_ips_enabled_flag= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "true" == "true" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: # This is needed so that converting clusters from GA to TP Mar 18 10:14:13 crc kubenswrapper[4733]: # will rollout control plane pods as well Mar 18 10:14:13 crc kubenswrapper[4733]: network_segmentation_enabled_flag= Mar 18 10:14:13 crc kubenswrapper[4733]: multi_network_enabled_flag= Mar 18 10:14:13 crc kubenswrapper[4733]: if [[ "true" == "true" ]]; then Mar 18 10:14:13 crc kubenswrapper[4733]: multi_network_enabled_flag="--enable-multi-network" Mar 18 10:14:13 crc kubenswrapper[4733]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 18 10:14:13 crc kubenswrapper[4733]: fi Mar 18 10:14:13 crc kubenswrapper[4733]: Mar 18 10:14:13 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 18 10:14:13 crc kubenswrapper[4733]: exec /usr/bin/ovnkube \ Mar 18 10:14:13 crc kubenswrapper[4733]: --enable-interconnect \ Mar 18 10:14:13 crc kubenswrapper[4733]: --init-cluster-manager "${K8S_NODE}" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 18 10:14:13 crc kubenswrapper[4733]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --metrics-bind-address "127.0.0.1:29108" \ Mar 18 10:14:13 crc kubenswrapper[4733]: --metrics-enable-pprof \ Mar 18 10:14:13 crc kubenswrapper[4733]: --metrics-enable-config-duration \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${ovn_v4_join_subnet_opt} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${ovn_v6_join_subnet_opt} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${dns_name_resolver_enabled_flag} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${persistent_ips_enabled_flag} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${multi_network_enabled_flag} \ Mar 18 10:14:13 crc kubenswrapper[4733]: ${network_segmentation_enabled_flag} Mar 18 10:14:13 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vg7hc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-spfjj_openshift-ovn-kubernetes(7d693a73-68c1-4595-bbcc-be97691b06fe): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.614999 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.615472 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: W0318 10:14:13.616110 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc85b0d4_15a5_4894_9f07_9aaeb28f63fa.slice/crio-f6b7bfe1b02ca90d8b984d1a9c50c920b6a4ec109f2b15d0d9e4be17d1064a45 WatchSource:0}: Error finding container f6b7bfe1b02ca90d8b984d1a9c50c920b6a4ec109f2b15d0d9e4be17d1064a45: Status 404 returned error can't find the container with id f6b7bfe1b02ca90d8b984d1a9c50c920b6a4ec109f2b15d0d9e4be17d1064a45 Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.616120 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.616174 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" podUID="7d693a73-68c1-4595-bbcc-be97691b06fe" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.617925 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpnv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.617954 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:13 crc kubenswrapper[4733]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 18 10:14:13 crc kubenswrapper[4733]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 18 10:14:13 crc kubenswrapper[4733]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ph8vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-g6j2q_openshift-multus(cc85b0d4-15a5-4894-9f07-9aaeb28f63fa): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:13 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.619219 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-g6j2q" podUID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.620346 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpnv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.621423 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.625144 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.638335 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.651949 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.651986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.651999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.652016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.652028 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.655370 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.670942 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.680144 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.695476 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.707024 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.723159 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.733614 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.748215 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.755272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.755332 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.755353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.755382 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.755401 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.759860 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.772323 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.787486 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.801046 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.812701 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.824354 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.835943 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.845075 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.856709 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.861819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.861856 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.861868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.861888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.861902 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.866088 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.883736 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.883932 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.883902848 +0000 UTC m=+94.375637173 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.883986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.884056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.884120 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.884171 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884172 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884251 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884300 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.884291989 +0000 UTC m=+94.376026314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884326 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.884316359 +0000 UTC m=+94.376050684 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884360 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884387 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884406 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884468 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.884448983 +0000 UTC m=+94.376183318 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884609 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.884656 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.884644489 +0000 UTC m=+94.376378814 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.887026 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.927571 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.964876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.964935 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.964946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.964963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.964973 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:13Z","lastTransitionTime":"2026-03-18T10:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.968646 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:13 crc kubenswrapper[4733]: I0318 10:14:13.985675 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.985868 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.985891 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.985905 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:13 crc kubenswrapper[4733]: E0318 10:14:13.985958 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:14.985941686 +0000 UTC m=+94.477676011 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.010264 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.050080 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.067498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.067542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.067555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.067572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.067586 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.106434 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.170153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.170218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.170231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.170248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.170261 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.272724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.273069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.273158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.273264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.273369 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.375579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.375616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.375624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.375647 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.375657 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.477967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.478018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.478030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.478049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.478066 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.578281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"44de9df63ec3ee4eae109c26660b6c1eb3f01d1f6242fd4f233e7ef6cfded6a1"} Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.582286 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpnv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582420 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582486 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" event={"ID":"7d693a73-68c1-4595-bbcc-be97691b06fe","Type":"ContainerStarted","Data":"ca6147b370c26e7fb7577a789dd3313109801f85318d443e4aaac91a1c487a78"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.582478 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.591907 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpnv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.592175 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:14 crc kubenswrapper[4733]: set -euo pipefail Mar 18 10:14:14 crc kubenswrapper[4733]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 18 10:14:14 crc kubenswrapper[4733]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 18 10:14:14 crc kubenswrapper[4733]: # As the secret mount is optional we must wait for the files to be present. Mar 18 10:14:14 crc kubenswrapper[4733]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 18 10:14:14 crc kubenswrapper[4733]: TS=$(date +%s) Mar 18 10:14:14 crc kubenswrapper[4733]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 18 10:14:14 crc kubenswrapper[4733]: HAS_LOGGED_INFO=0 Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: log_missing_certs(){ Mar 18 10:14:14 crc kubenswrapper[4733]: CUR_TS=$(date +%s) Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 18 10:14:14 crc kubenswrapper[4733]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 18 10:14:14 crc kubenswrapper[4733]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 18 10:14:14 crc kubenswrapper[4733]: HAS_LOGGED_INFO=1 Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: } Mar 18 10:14:14 crc kubenswrapper[4733]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 18 10:14:14 crc kubenswrapper[4733]: log_missing_certs Mar 18 10:14:14 crc kubenswrapper[4733]: sleep 5 Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 18 10:14:14 crc kubenswrapper[4733]: exec /usr/bin/kube-rbac-proxy \ Mar 18 10:14:14 crc kubenswrapper[4733]: --logtostderr \ Mar 18 10:14:14 crc kubenswrapper[4733]: --secure-listen-address=:9108 \ Mar 18 10:14:14 crc kubenswrapper[4733]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 18 10:14:14 crc kubenswrapper[4733]: --upstream=http://127.0.0.1:29108/ \ Mar 18 10:14:14 crc kubenswrapper[4733]: --tls-private-key-file=${TLS_PK} \ Mar 18 10:14:14 crc kubenswrapper[4733]: --tls-cert-file=${TLS_CERT} Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vg7hc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-spfjj_openshift-ovn-kubernetes(7d693a73-68c1-4595-bbcc-be97691b06fe): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.593206 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hsk58" event={"ID":"c2c181c8-3361-40a2-afc5-a677e0ab4ecd","Type":"ContainerStarted","Data":"16909e99ed8cac62104f4c439d3da9d1b3ffb5e99fada86f18c770cdbb02a00a"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.593250 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.593754 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.597722 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:14 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v4_join_subnet_opt= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v6_join_subnet_opt= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v4_transit_switch_subnet_opt= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v6_transit_switch_subnet_opt= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "" != "" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: dns_name_resolver_enabled_flag= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "false" == "true" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: persistent_ips_enabled_flag= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "true" == "true" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: # This is needed so that converting clusters from GA to TP Mar 18 10:14:14 crc kubenswrapper[4733]: # will rollout control plane pods as well Mar 18 10:14:14 crc kubenswrapper[4733]: network_segmentation_enabled_flag= Mar 18 10:14:14 crc kubenswrapper[4733]: multi_network_enabled_flag= Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "true" == "true" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: multi_network_enabled_flag="--enable-multi-network" Mar 18 10:14:14 crc kubenswrapper[4733]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 18 10:14:14 crc kubenswrapper[4733]: exec /usr/bin/ovnkube \ Mar 18 10:14:14 crc kubenswrapper[4733]: --enable-interconnect \ Mar 18 10:14:14 crc kubenswrapper[4733]: --init-cluster-manager "${K8S_NODE}" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 18 10:14:14 crc kubenswrapper[4733]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --metrics-bind-address "127.0.0.1:29108" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --metrics-enable-pprof \ Mar 18 10:14:14 crc kubenswrapper[4733]: --metrics-enable-config-duration \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${ovn_v4_join_subnet_opt} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${ovn_v6_join_subnet_opt} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${dns_name_resolver_enabled_flag} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${persistent_ips_enabled_flag} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${multi_network_enabled_flag} \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${network_segmentation_enabled_flag} Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vg7hc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-spfjj_openshift-ovn-kubernetes(7d693a73-68c1-4595-bbcc-be97691b06fe): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.597956 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 18 10:14:14 crc kubenswrapper[4733]: set -uo pipefail Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 18 10:14:14 crc kubenswrapper[4733]: HOSTS_FILE="/etc/hosts" Mar 18 10:14:14 crc kubenswrapper[4733]: TEMP_FILE="/etc/hosts.tmp" Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: # Make a temporary file with the old hosts file's attributes. Mar 18 10:14:14 crc kubenswrapper[4733]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 18 10:14:14 crc kubenswrapper[4733]: echo "Failed to preserve hosts file. Exiting." Mar 18 10:14:14 crc kubenswrapper[4733]: exit 1 Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: while true; do Mar 18 10:14:14 crc kubenswrapper[4733]: declare -A svc_ips Mar 18 10:14:14 crc kubenswrapper[4733]: for svc in "${services[@]}"; do Mar 18 10:14:14 crc kubenswrapper[4733]: # Fetch service IP from cluster dns if present. We make several tries Mar 18 10:14:14 crc kubenswrapper[4733]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 18 10:14:14 crc kubenswrapper[4733]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 18 10:14:14 crc kubenswrapper[4733]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 18 10:14:14 crc kubenswrapper[4733]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:14 crc kubenswrapper[4733]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:14 crc kubenswrapper[4733]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 18 10:14:14 crc kubenswrapper[4733]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 18 10:14:14 crc kubenswrapper[4733]: for i in ${!cmds[*]} Mar 18 10:14:14 crc kubenswrapper[4733]: do Mar 18 10:14:14 crc kubenswrapper[4733]: ips=($(eval "${cmds[i]}")) Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: svc_ips["${svc}"]="${ips[@]}" Mar 18 10:14:14 crc kubenswrapper[4733]: break Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: # Update /etc/hosts only if we get valid service IPs Mar 18 10:14:14 crc kubenswrapper[4733]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 18 10:14:14 crc kubenswrapper[4733]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 18 10:14:14 crc kubenswrapper[4733]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 18 10:14:14 crc kubenswrapper[4733]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 18 10:14:14 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:14 crc kubenswrapper[4733]: continue Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: # Append resolver entries for services Mar 18 10:14:14 crc kubenswrapper[4733]: rc=0 Mar 18 10:14:14 crc kubenswrapper[4733]: for svc in "${!svc_ips[@]}"; do Mar 18 10:14:14 crc kubenswrapper[4733]: for ip in ${svc_ips[${svc}]}; do Mar 18 10:14:14 crc kubenswrapper[4733]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ $rc -ne 0 ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:14 crc kubenswrapper[4733]: continue Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 18 10:14:14 crc kubenswrapper[4733]: # Replace /etc/hosts with our modified version if needed Mar 18 10:14:14 crc kubenswrapper[4733]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 18 10:14:14 crc kubenswrapper[4733]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: sleep 60 & wait Mar 18 10:14:14 crc kubenswrapper[4733]: unset svc_ips Mar 18 10:14:14 crc kubenswrapper[4733]: done Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-httph,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-hsk58_openshift-dns(c2c181c8-3361-40a2-afc5-a677e0ab4ecd): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.598890 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" podUID="7d693a73-68c1-4595-bbcc-be97691b06fe" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.599111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-hsk58" podUID="c2c181c8-3361-40a2-afc5-a677e0ab4ecd" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.599958 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerStarted","Data":"ff1d93cef35a316c51b4c062176258993d12b241a2a73aaf584a089109ebf4c1"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.606120 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"082c29e3548c80a38ab2a902a222a7bd38f378ff323b000dc303bf38651a6b0f"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.607400 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.607619 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwk4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-t28sh_openshift-multus(0f82588a-9dbd-4c55-8cfc-f96e57fa58b9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.608910 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-t28sh" podUID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.610970 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.612125 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7fb0376f2eec0926827408777c433a08963622c37f222da5c978e502f9dbcbfd"} Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.612816 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.614687 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:14 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 18 10:14:14 crc kubenswrapper[4733]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 18 10:14:14 crc kubenswrapper[4733]: ho_enable="--enable-hybrid-overlay" Mar 18 10:14:14 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 18 10:14:14 crc kubenswrapper[4733]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 18 10:14:14 crc kubenswrapper[4733]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 18 10:14:14 crc kubenswrapper[4733]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 10:14:14 crc kubenswrapper[4733]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --webhook-host=127.0.0.1 \ Mar 18 10:14:14 crc kubenswrapper[4733]: --webhook-port=9743 \ Mar 18 10:14:14 crc kubenswrapper[4733]: ${ho_enable} \ Mar 18 10:14:14 crc kubenswrapper[4733]: --enable-interconnect \ Mar 18 10:14:14 crc kubenswrapper[4733]: --disable-approver \ Mar 18 10:14:14 crc kubenswrapper[4733]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --wait-for-kubernetes-api=200s \ Mar 18 10:14:14 crc kubenswrapper[4733]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --loglevel="${LOGLEVEL}" Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.614839 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerStarted","Data":"f6b7bfe1b02ca90d8b984d1a9c50c920b6a4ec109f2b15d0d9e4be17d1064a45"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.616593 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.616985 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 18 10:14:14 crc kubenswrapper[4733]: apiVersion: v1 Mar 18 10:14:14 crc kubenswrapper[4733]: clusters: Mar 18 10:14:14 crc kubenswrapper[4733]: - cluster: Mar 18 10:14:14 crc kubenswrapper[4733]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 18 10:14:14 crc kubenswrapper[4733]: server: https://api-int.crc.testing:6443 Mar 18 10:14:14 crc kubenswrapper[4733]: name: default-cluster Mar 18 10:14:14 crc kubenswrapper[4733]: contexts: Mar 18 10:14:14 crc kubenswrapper[4733]: - context: Mar 18 10:14:14 crc kubenswrapper[4733]: cluster: default-cluster Mar 18 10:14:14 crc kubenswrapper[4733]: namespace: default Mar 18 10:14:14 crc kubenswrapper[4733]: user: default-auth Mar 18 10:14:14 crc kubenswrapper[4733]: name: default-context Mar 18 10:14:14 crc kubenswrapper[4733]: current-context: default-context Mar 18 10:14:14 crc kubenswrapper[4733]: kind: Config Mar 18 10:14:14 crc kubenswrapper[4733]: preferences: {} Mar 18 10:14:14 crc kubenswrapper[4733]: users: Mar 18 10:14:14 crc kubenswrapper[4733]: - name: default-auth Mar 18 10:14:14 crc kubenswrapper[4733]: user: Mar 18 10:14:14 crc kubenswrapper[4733]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 18 10:14:14 crc kubenswrapper[4733]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 18 10:14:14 crc kubenswrapper[4733]: EOF Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zqxdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.618124 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 10:14:14 crc kubenswrapper[4733]: if [[ -f "/env/_master" ]]; then Mar 18 10:14:14 crc kubenswrapper[4733]: set -o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: source "/env/_master" Mar 18 10:14:14 crc kubenswrapper[4733]: set +o allexport Mar 18 10:14:14 crc kubenswrapper[4733]: fi Mar 18 10:14:14 crc kubenswrapper[4733]: Mar 18 10:14:14 crc kubenswrapper[4733]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 18 10:14:14 crc kubenswrapper[4733]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 10:14:14 crc kubenswrapper[4733]: --disable-webhook \ Mar 18 10:14:14 crc kubenswrapper[4733]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 18 10:14:14 crc kubenswrapper[4733]: --loglevel="${LOGLEVEL}" Mar 18 10:14:14 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.618168 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.618362 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:14:14 crc kubenswrapper[4733]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 18 10:14:14 crc kubenswrapper[4733]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 18 10:14:14 crc kubenswrapper[4733]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ph8vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-g6j2q_openshift-multus(cc85b0d4-15a5-4894-9f07-9aaeb28f63fa): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 10:14:14 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.619377 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.619547 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-g6j2q" podUID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.624654 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.636686 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.645595 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.655118 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.664865 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.676893 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.685546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.685601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.685611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.685629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.685642 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.686037 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.692960 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.708601 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.724396 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.733668 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.740735 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.750370 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.767564 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.807559 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.819330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.819394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.819404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.819423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.819436 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.849295 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.887457 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.918173 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.918380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.918423 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.918445 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918503 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:16.918483803 +0000 UTC m=+96.410218128 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.918555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918619 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918679 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918713 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:16.918705499 +0000 UTC m=+96.410439825 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918742 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:16.91871922 +0000 UTC m=+96.410453545 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918629 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918774 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918786 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918812 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:16.918805922 +0000 UTC m=+96.410540247 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.918834 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: E0318 10:14:14.919030 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:16.918973547 +0000 UTC m=+96.410708062 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.923105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.923160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.923183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.923253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.923278 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:14Z","lastTransitionTime":"2026-03-18T10:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.934220 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:14 crc kubenswrapper[4733]: I0318 10:14:14.967108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.006555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.019772 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.020148 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.020254 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.020288 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.020409 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:17.020374287 +0000 UTC m=+96.512108772 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.026660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.026718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.026735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.026760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.026778 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.050959 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.088463 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.129583 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.130234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.130322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.130353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.130390 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.130417 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.172948 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.175437 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.175583 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.175721 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.175790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.175890 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.175980 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.176004 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:15 crc kubenswrapper[4733]: E0318 10:14:15.176329 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.181601 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.184512 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.186138 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.188778 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.190122 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.192336 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.193736 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.194628 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.195424 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.196353 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.197239 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.198026 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.199274 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.200264 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.201321 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.202075 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.205894 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.206929 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.208050 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.208879 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.209744 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.211360 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.212340 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.213744 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.214753 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.215532 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.217098 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.218692 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.219480 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.220436 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.221794 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.222581 4733 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.222795 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.225540 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.226076 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.226803 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.228933 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.229604 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.230780 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.231462 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.232864 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.233415 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234176 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.234469 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.235054 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.235283 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.236319 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.236882 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.237965 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.238580 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.239702 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.240319 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.241231 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.241863 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.242433 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.243423 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.243921 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.267707 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.317298 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.338450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.338610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.338678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.338792 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.338902 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.442633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.442707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.442726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.442755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.442775 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.545932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.545998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.546020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.546052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.546077 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.648947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.649042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.649057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.649077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.649093 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.755433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.755468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.755479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.755496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.755510 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.858007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.858055 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.858067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.858083 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.858094 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.961143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.961211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.961220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.961241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:15 crc kubenswrapper[4733]: I0318 10:14:15.961253 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:15Z","lastTransitionTime":"2026-03-18T10:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.064311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.064383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.064405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.064435 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.064489 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.166792 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.166875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.166896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.166927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.166950 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.269844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.269897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.269906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.269944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.269956 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.372912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.372960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.372969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.372985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.372996 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.476134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.476169 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.476177 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.476207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.476217 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.578306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.578337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.578345 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.578360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.578370 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.681815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.681880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.681893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.681916 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.681928 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.785863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.785909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.785926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.785947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.785964 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.889379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.889448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.889470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.889498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.889519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.942161 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.942383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942423 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:20.942381251 +0000 UTC m=+100.434115616 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.942462 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.942508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.942568 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942691 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942703 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942755 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:20.942739751 +0000 UTC m=+100.434474116 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942752 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942756 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942906 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942713 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.942849 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:20.942819763 +0000 UTC m=+100.434554128 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.943087 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:20.94305852 +0000 UTC m=+100.434792875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:16 crc kubenswrapper[4733]: E0318 10:14:16.943116 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:20.943102061 +0000 UTC m=+100.434836416 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.993657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.993732 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.993755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.993784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:16 crc kubenswrapper[4733]: I0318 10:14:16.993807 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:16Z","lastTransitionTime":"2026-03-18T10:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.043777 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.043993 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.044027 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.044040 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.044106 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:21.04408664 +0000 UTC m=+100.535820965 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.096372 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.096432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.096450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.096508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.096528 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.174720 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.174816 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.174883 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.175010 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.175105 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.176363 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.176549 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:17 crc kubenswrapper[4733]: E0318 10:14:17.176916 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.200112 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.200333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.200394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.200421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.200440 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.303124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.303213 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.303233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.303264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.303289 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.406724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.406776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.406792 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.406811 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.406825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.510143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.510210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.510227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.510249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.510266 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.613378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.613434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.613448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.613478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.613495 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.715984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.716022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.716034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.716052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.716065 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.819095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.819172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.819242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.819273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.819293 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.922248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.922317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.922343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.922373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:17 crc kubenswrapper[4733]: I0318 10:14:17.922402 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:17Z","lastTransitionTime":"2026-03-18T10:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.025999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.026063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.026081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.026109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.026130 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.130608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.130708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.130737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.130782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.130825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.234687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.234746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.234756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.234773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.234783 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.338257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.338362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.338386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.338413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.338431 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.441006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.441084 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.441108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.441140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.441163 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.544317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.544398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.544422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.544453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.544475 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.647717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.647769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.647786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.647808 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.647825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.750702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.750750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.750762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.750781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.750797 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.853609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.853647 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.853655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.853669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.853681 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.956593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.956766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.956857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.956887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:18 crc kubenswrapper[4733]: I0318 10:14:18.956960 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:18Z","lastTransitionTime":"2026-03-18T10:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.059941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.059996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.060008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.060027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.060041 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.163588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.163737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.163754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.163780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.163798 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.175476 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.175568 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.175612 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.175851 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:19 crc kubenswrapper[4733]: E0318 10:14:19.175873 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:19 crc kubenswrapper[4733]: E0318 10:14:19.176003 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:19 crc kubenswrapper[4733]: E0318 10:14:19.176137 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:19 crc kubenswrapper[4733]: E0318 10:14:19.176300 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.267227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.267274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.267284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.267302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.267313 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.370642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.370726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.370746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.370777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.370798 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.474401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.474491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.474512 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.474545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.474566 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.579305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.579383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.579411 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.579443 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.579461 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.683409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.683495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.683516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.683549 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.683571 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.786810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.786902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.786921 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.786951 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.786972 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.890751 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.890869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.890891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.890919 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.890938 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.994569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.994654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.994680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.994716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:19 crc kubenswrapper[4733]: I0318 10:14:19.994741 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:19Z","lastTransitionTime":"2026-03-18T10:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.098403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.098478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.098496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.098525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.098544 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.203246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.203322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.203350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.203383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.203407 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.307093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.307249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.307277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.307309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.307332 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.418367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.418888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.419106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.420113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.420253 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.525822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.525871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.525893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.525923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.525942 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.630276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.630418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.630438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.630468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.630486 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.734010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.734087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.734113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.734145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.734168 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.838032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.838109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.838130 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.838160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.838183 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.941559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.941650 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.941677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.941713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.941741 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:20Z","lastTransitionTime":"2026-03-18T10:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.993121 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993401 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:28.993357173 +0000 UTC m=+108.485091538 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.993471 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.993534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.993619 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:20 crc kubenswrapper[4733]: I0318 10:14:20.993679 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993706 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993792 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:28.993767385 +0000 UTC m=+108.485501740 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993814 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993909 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:28.993878618 +0000 UTC m=+108.485612983 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.993933 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.994031 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:28.994000021 +0000 UTC m=+108.485734386 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.994026 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.994090 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.994119 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:20 crc kubenswrapper[4733]: E0318 10:14:20.994180 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:28.994163716 +0000 UTC m=+108.485898071 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.044803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.044901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.044931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.044969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.044994 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.094951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.095245 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.095290 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.095306 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.095379 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:29.09535735 +0000 UTC m=+108.587091685 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.148229 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.148305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.148323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.148358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.148382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.174526 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.174556 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.174726 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.174904 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.174973 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.175220 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.175370 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.175530 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.190690 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.200535 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.210268 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.222404 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.237099 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.249892 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.252326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.252403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.252425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.252455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.252481 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.275635 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.295664 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.311983 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.324261 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.340482 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.357573 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.357755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.357747 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.357793 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.358023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.358044 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.370613 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.387606 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.399762 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.461643 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.461696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.461707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.461721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.461731 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.527914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.527982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.528004 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.528029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.528048 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.541056 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.547359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.547424 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.547442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.547473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.547495 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.599008 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.604599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.604671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.604692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.604768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.604825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.621307 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.626788 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.626874 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.626894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.626948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.626965 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.643320 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.649950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.650024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.650042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.650072 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.650091 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.665972 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:21 crc kubenswrapper[4733]: E0318 10:14:21.666144 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.668685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.668805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.668896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.668970 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.669000 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.773101 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.773177 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.773226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.773248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.773266 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.877000 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.877071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.877089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.877119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.877139 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.981693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.981762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.981779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.981805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.981822 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:21Z","lastTransitionTime":"2026-03-18T10:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:21 crc kubenswrapper[4733]: I0318 10:14:21.982230 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.086448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.086545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.086566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.086599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.086624 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.192837 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.193627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.193700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.193738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.193770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.193796 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.196798 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.197021 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.298819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.298859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.298872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.298891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.298905 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.402249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.402685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.402695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.402709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.402726 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.505917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.506053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.506081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.506145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.506173 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.610094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.610134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.610149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.610170 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.610181 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.641157 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.643800 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.660405 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.674379 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.701381 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.712769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.712820 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.712834 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.712852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.712872 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.714322 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.726105 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.740877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.753606 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.765238 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.795512 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.809822 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.815458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.815523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.815551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.815588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.815618 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.821905 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.836297 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.848640 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.862610 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.875944 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.892327 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.905669 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.918234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.918327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.918396 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.918464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:22 crc kubenswrapper[4733]: I0318 10:14:22.918523 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:22Z","lastTransitionTime":"2026-03-18T10:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.021543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.022386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.022503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.022593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.022684 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.126231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.126580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.126868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.126978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.127101 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.175480 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.175505 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.175543 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.175479 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:23 crc kubenswrapper[4733]: E0318 10:14:23.175632 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:23 crc kubenswrapper[4733]: E0318 10:14:23.175714 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:23 crc kubenswrapper[4733]: E0318 10:14:23.175785 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:23 crc kubenswrapper[4733]: E0318 10:14:23.175862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.230878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.231271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.231419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.231560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.231708 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.335431 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.335846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.335933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.336019 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.336096 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.439369 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.439429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.439445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.439471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.439486 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.542620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.542666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.542680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.542704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.542719 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.645954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.646402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.646507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.646607 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.646694 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.647038 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.750096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.750610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.750838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.751054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.751267 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.855263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.855317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.855330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.855350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.855363 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.958767 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.958834 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.958852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.958881 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:23 crc kubenswrapper[4733]: I0318 10:14:23.958901 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:23Z","lastTransitionTime":"2026-03-18T10:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.062826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.062890 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.062909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.062941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.062973 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.167110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.167465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.167536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.167607 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.167669 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.270492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.270779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.271013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.271183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.271298 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.375057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.375400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.375493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.375590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.375665 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.478902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.478958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.478973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.478996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.479018 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.582713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.583013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.583168 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.583312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.583428 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.686227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.686284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.686300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.686328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.686342 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.790516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.790611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.790638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.790673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.790697 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.894109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.894566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.895030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.895427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:24 crc kubenswrapper[4733]: I0318 10:14:24.895742 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:24Z","lastTransitionTime":"2026-03-18T10:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:24.999901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.000740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.000981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.001183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.001473 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.104999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.106381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.106436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.106466 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.106485 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.175369 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.175429 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.175507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:25 crc kubenswrapper[4733]: E0318 10:14:25.175504 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.175429 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:25 crc kubenswrapper[4733]: E0318 10:14:25.175697 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:25 crc kubenswrapper[4733]: E0318 10:14:25.175795 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:25 crc kubenswrapper[4733]: E0318 10:14:25.175885 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.209723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.209785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.209807 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.209833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.209853 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.312646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.312716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.312739 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.312787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.312809 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.416060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.416129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.416149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.416177 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.416227 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.519796 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.519868 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.519892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.519919 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.519940 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.623046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.623125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.623145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.623256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.623279 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.726866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.726915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.726932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.726957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.726973 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.830297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.830388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.830400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.830415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.830425 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.933587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.934079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.934249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.934414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:25 crc kubenswrapper[4733]: I0318 10:14:25.934557 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:25Z","lastTransitionTime":"2026-03-18T10:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.038128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.038172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.038201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.038221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.038234 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.141412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.141484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.141510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.141576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.141599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.245939 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.246017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.246039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.246073 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.246096 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.349251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.349480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.349492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.349511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.349526 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.452791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.453471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.453545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.453643 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.453673 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.557053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.557101 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.557112 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.557132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.557149 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerStarted","Data":"cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659182 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.659240 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.662456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xfvfl" event={"ID":"bb58b528-9013-4fab-9747-60bb6ff1bc1f","Type":"ContainerStarted","Data":"fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.675134 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.690931 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.702334 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.715168 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.725483 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.733709 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.761309 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.767699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.767740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.767752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.767773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.767786 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.778228 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.790625 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.804822 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.826347 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.838667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870808 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870871 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.870839 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.883229 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.895252 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.912435 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.931037 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.946506 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.959855 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.974720 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.975164 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.975217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.975230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.975252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.975266 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:26Z","lastTransitionTime":"2026-03-18T10:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:26 crc kubenswrapper[4733]: I0318 10:14:26.992958 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.004596 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.014831 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.037085 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.066054 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.076673 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.078591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.078637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.078648 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.078670 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.078680 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.094754 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.106156 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.120555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.132947 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.145748 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.156859 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.167453 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.175417 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.175471 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:27 crc kubenswrapper[4733]: E0318 10:14:27.175614 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.175682 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:27 crc kubenswrapper[4733]: E0318 10:14:27.175730 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:27 crc kubenswrapper[4733]: E0318 10:14:27.175946 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.176055 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:27 crc kubenswrapper[4733]: E0318 10:14:27.176226 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.181003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.181714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.182461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.182661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.182798 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.181373 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.287602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.287664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.287676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.287695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.287712 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.390960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.391016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.391028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.391048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.391063 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.493682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.493760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.493832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.493871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.493895 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.598370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.598447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.598467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.598492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.598513 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.701930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.702004 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.702022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.702047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.702067 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.805351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.805422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.805439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.805464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.805483 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.909394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.909480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.909510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.909578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:27 crc kubenswrapper[4733]: I0318 10:14:27.909603 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:27Z","lastTransitionTime":"2026-03-18T10:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.012737 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.012818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.012843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.012876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.012897 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.116367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.116451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.116471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.116502 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.116525 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.227037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.227145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.227264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.227288 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.227310 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.331682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.331716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.331725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.331741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.331753 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.434659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.434702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.434713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.434727 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.434737 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.537938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.537992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.538036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.538060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.538073 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.641702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.641766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.641785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.641812 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.641832 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.678232 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.678331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.681291 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hsk58" event={"ID":"c2c181c8-3361-40a2-afc5-a677e0ab4ecd","Type":"ContainerStarted","Data":"a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.685447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerStarted","Data":"f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.688327 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.688384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.704603 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.721594 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.740511 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.746003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.746125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.746152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.746184 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.746273 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.756132 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.776737 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.791374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.803762 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.814592 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.830508 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.845913 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.848530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.848562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.848572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.848592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.848604 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.861120 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.872806 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.883047 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.895632 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.906299 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.929505 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:28Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.949116 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:28Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.951312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.951360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.951374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.951391 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.951402 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:28Z","lastTransitionTime":"2026-03-18T10:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.978928 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:28Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:28 crc kubenswrapper[4733]: I0318 10:14:28.993801 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:28Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.017805 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.038366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055494 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055645 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.055885 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.070278 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.084993 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.089570 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.089763 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.089817 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.089876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.089904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090033 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090117 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.090093874 +0000 UTC m=+124.581828219 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090171 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090239 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090267 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090236 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090361 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.090323781 +0000 UTC m=+124.582058146 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090416 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.090390923 +0000 UTC m=+124.582125438 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090544 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.090744 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.090729672 +0000 UTC m=+124.582464227 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.091023 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.090987409 +0000 UTC m=+124.582721774 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.103715 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.122478 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.139693 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.151014 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.157756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.157885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.157963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.158047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.158170 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.159761 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.170943 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.175375 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.175380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.175432 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.175659 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.175792 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.176774 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.176835 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.176912 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.184224 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.190595 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.191475 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.191541 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.191568 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:29 crc kubenswrapper[4733]: E0318 10:14:29.191651 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:14:45.191621658 +0000 UTC m=+124.683356053 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.206623 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.220237 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.241524 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.262635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.262676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.262688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.262705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.262716 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.365278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.365336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.365351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.365388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.365406 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.468081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.468132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.468149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.468172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.468224 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.572226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.572288 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.572314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.572341 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.572362 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.675303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.675815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.675834 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.675860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.675879 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.694758 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642" exitCode=0 Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.694868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.699715 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" exitCode=0 Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.699772 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.720406 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.740398 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.757687 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.777898 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.779593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.779622 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.779633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.779652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.779663 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.794063 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.806623 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.825520 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.843961 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.859096 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.872564 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.882649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.882693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.882704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.882720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.882731 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.896487 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.933751 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.952037 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.970180 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.986972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.987343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.987354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.987369 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.987379 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:29Z","lastTransitionTime":"2026-03-18T10:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:29 crc kubenswrapper[4733]: I0318 10:14:29.987225 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:29Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.008710 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.024053 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.039859 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.053325 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.065960 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.078763 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.089683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.089722 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.089734 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.089752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.089765 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.103046 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.114575 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.140497 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.166200 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.179711 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.192479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.192680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.192771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.192859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.192921 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.196671 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.214622 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.239955 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.254985 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.276752 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.293816 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.295084 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.295129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.295140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.295155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.295165 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.309006 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.328888 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.398911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.398954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.398967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.398992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.399004 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.502688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.502726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.502767 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.502805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.502819 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.605279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.605307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.605316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.605331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.605339 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.710463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.710503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.710547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.710567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.710580 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.712771 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.719713 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc" exitCode=0 Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.719796 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727272 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727327 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727363 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.727373 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.730158 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.733048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.734938 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" event={"ID":"7d693a73-68c1-4595-bbcc-be97691b06fe","Type":"ContainerStarted","Data":"e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.734970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" event={"ID":"7d693a73-68c1-4595-bbcc-be97691b06fe","Type":"ContainerStarted","Data":"cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.740333 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.758430 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.773872 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.789932 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.802503 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.812715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.812901 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.812999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.813088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.813169 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.819253 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.831937 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.849825 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.863764 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.878909 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.894913 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.912207 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.916730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.916772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.916784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.916803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.916815 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:30Z","lastTransitionTime":"2026-03-18T10:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.926558 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.938172 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.962921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.981624 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:30 crc kubenswrapper[4733]: I0318 10:14:30.994444 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:30Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.007216 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.020140 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.021110 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.021265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.021353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.021444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.021532 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.035992 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.054833 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.067561 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.098647 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.121928 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.124519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.124593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.124612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.124641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.124662 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.137343 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.154381 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.173431 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.174771 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.174821 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.174885 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.174914 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.174991 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.175138 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.175351 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.175458 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.217374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.227375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.227412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.227422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.227439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.227449 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.268077 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.289031 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.310150 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.325039 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.329719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.329748 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.329757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.329776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.329785 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.337864 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.350352 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.359950 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.370421 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.383394 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.399095 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.412902 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.425485 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.432556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.432595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.432605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.432624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.432639 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.440957 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.458625 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.471737 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.492658 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.521911 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.532880 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.535499 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.535539 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.535549 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.535566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.535577 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.550555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.564870 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.582759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.600719 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.639077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.639116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.639125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.639138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.639147 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.744394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.744445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.744460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.744480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.744495 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.745604 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637" exitCode=0 Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.745685 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.771327 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.786652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.786688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.786698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.786714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.786725 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.790769 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.806270 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.811441 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.813525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.813564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.813583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.813606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.813626 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.837766 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.846869 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.848445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.848479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.848489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.848508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.848523 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.871721 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.874521 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.879985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.880012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.880022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.880039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.880050 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.887877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.895481 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.899267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.899307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.899323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.899343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.899366 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.909236 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.923043 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: E0318 10:14:31.923252 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.925725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.925761 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.925775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.925795 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.925810 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:31Z","lastTransitionTime":"2026-03-18T10:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.929805 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.948254 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.962412 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:31 crc kubenswrapper[4733]: I0318 10:14:31.991371 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.007563 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.028327 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.029244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.029291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.029302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.029326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.029338 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.042401 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.063142 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.095500 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.132582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.132669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.132698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.132736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.132761 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.136522 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.242071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.242150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.242171 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.242231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.242251 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.345709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.345803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.345831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.345866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.345888 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.450625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.450705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.450726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.450762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.450781 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.554956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.555865 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.555909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.555940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.555961 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.658908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.658986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.659053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.659091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.659114 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.754830 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256" exitCode=0 Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.754925 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.764232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.764445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.764573 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.764733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.764860 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.777150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.786689 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.808914 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.832290 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.854247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.868531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.868605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.868632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.868669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.868699 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.888543 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.914217 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.933148 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.953660 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.971865 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.971920 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.971939 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.971968 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.971987 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:32Z","lastTransitionTime":"2026-03-18T10:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:32 crc kubenswrapper[4733]: I0318 10:14:32.979301 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.000454 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:32Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.035868 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.049109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.072749 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.076622 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.076685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.076703 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.076731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.076751 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.088921 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.104466 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.121702 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.143223 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.175522 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:33 crc kubenswrapper[4733]: E0318 10:14:33.175727 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.176408 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.176447 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.176846 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:33 crc kubenswrapper[4733]: E0318 10:14:33.176857 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:33 crc kubenswrapper[4733]: E0318 10:14:33.176961 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:33 crc kubenswrapper[4733]: E0318 10:14:33.177121 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.179014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.179033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.179041 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.179053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.179063 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.282508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.282584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.282608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.282632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.282649 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.385620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.385667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.385677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.385699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.385710 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.489149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.489233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.489244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.489272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.489283 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.592813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.592910 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.592934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.592969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.593001 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.696374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.696421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.696430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.696448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.696458 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.789067 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80" exitCode=0 Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.789149 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.802844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.803582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.803605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.803635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.803656 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.819018 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.838905 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.897857 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.923843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.923885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.923895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.923917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.923931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:33Z","lastTransitionTime":"2026-03-18T10:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.930756 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.946316 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.966941 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.977009 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:33 crc kubenswrapper[4733]: I0318 10:14:33.989979 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.002413 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.017860 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.026117 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.026147 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.026160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.026181 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.026234 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.031130 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.049681 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.061922 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.076719 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.088696 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.100873 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.113159 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.129005 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.129063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.129082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.129108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.129129 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.232497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.232542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.232552 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.232568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.232577 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.336092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.336175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.336261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.336301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.336328 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.439374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.439438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.439450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.439469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.439531 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.542755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.542813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.542826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.542880 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.542893 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.646595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.646665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.646687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.646713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.646731 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.749915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.749966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.749983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.750011 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.750031 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.800649 4733 generic.go:334] "Generic (PLEG): container finished" podID="0f82588a-9dbd-4c55-8cfc-f96e57fa58b9" containerID="9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9" exitCode=0 Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.800710 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerDied","Data":"9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.827301 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.846012 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.854837 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.854891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.854905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.854957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.854971 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.867965 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.884583 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.900810 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.917719 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.931059 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.946555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.960646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.960683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.960692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.960709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.960720 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:34Z","lastTransitionTime":"2026-03-18T10:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.963247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:34 crc kubenswrapper[4733]: I0318 10:14:34.976593 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:34Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.005237 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.022614 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.042661 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.061717 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.063782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.063815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.063827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.063847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.063862 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.083340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.098518 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.120694 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.167335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.167409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.167421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.167442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.167454 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.175219 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.175248 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.175263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:35 crc kubenswrapper[4733]: E0318 10:14:35.175344 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:35 crc kubenswrapper[4733]: E0318 10:14:35.175527 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.175589 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:35 crc kubenswrapper[4733]: E0318 10:14:35.176330 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:35 crc kubenswrapper[4733]: E0318 10:14:35.176704 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.194984 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.271026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.271087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.271109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.271133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.271151 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.373972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.374012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.374023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.374038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.374075 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.476967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.477022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.477036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.477061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.477076 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.581452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.581997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.582023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.582058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.582081 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.684526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.684562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.684576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.684592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.684605 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.788515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.788616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.788634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.788658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.788679 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.812137 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.812667 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.812727 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.821961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" event={"ID":"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9","Type":"ContainerStarted","Data":"14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.835535 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.852649 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.863918 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.886338 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.892392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.892460 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.892478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.892505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.892523 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.910315 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.931265 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.955548 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.974792 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.992442 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:35Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.995520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.995589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.995608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.995633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:35 crc kubenswrapper[4733]: I0318 10:14:35.995654 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:35Z","lastTransitionTime":"2026-03-18T10:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.012493 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.036839 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.067786 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.086737 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.098379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.098451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.098471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.098499 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.098530 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.106419 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.129563 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.147764 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.167998 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.188234 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.203069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.203146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.203167 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.203223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.203470 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.231617 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.257090 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.277729 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.306674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.306746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.306767 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.306794 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.306815 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.312788 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.331901 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.357014 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.378608 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.405348 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.409719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.409797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.409819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.409848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.409868 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.431926 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.453773 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.465900 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.477170 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.494705 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512213 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512258 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.512175 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.521054 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.531485 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.552815 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.570603 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.583995 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.613375 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.614319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.614365 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.614381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.614406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.614422 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.649042 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.668619 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.689172 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.703434 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.716612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.716659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.716672 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.716691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.716702 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.726806 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.737661 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.775810 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.797403 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.813680 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.818998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.819066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.819079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.819096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.819106 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.825221 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.831222 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.847109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.853973 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.863109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.874951 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.888870 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.902009 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.915698 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.921032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.921086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.921103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.921125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.921140 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:36Z","lastTransitionTime":"2026-03-18T10:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.928805 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.949326 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.963593 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:36 crc kubenswrapper[4733]: I0318 10:14:36.975105 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:36Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.010281 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.026276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.026335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.026348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.026367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.026381 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.036927 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.050981 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.064932 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.080734 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.091421 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.112945 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.125272 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.128759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.128810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.128819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.128837 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.128850 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.138712 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.149178 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.162326 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.176499 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.176579 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.176526 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.176652 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:37 crc kubenswrapper[4733]: E0318 10:14:37.176643 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:37 crc kubenswrapper[4733]: E0318 10:14:37.176756 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:37 crc kubenswrapper[4733]: E0318 10:14:37.176864 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:37 crc kubenswrapper[4733]: E0318 10:14:37.176930 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.180971 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.194876 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.207239 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.221996 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.230911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.230940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.230948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.230961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.230970 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.235154 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.334091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.334142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.334152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.334166 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.334178 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.437093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.437406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.437472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.437534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.437598 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.540012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.540049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.540060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.540078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.540086 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.642917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.642958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.642969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.642983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.642995 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.746137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.746175 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.746203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.746224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.746235 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.830262 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/0.log" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.833511 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96" exitCode=1 Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.833584 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.834675 4733 scope.go:117] "RemoveContainer" containerID="f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.848524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.848563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.848576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.848593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.848608 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.858759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.874514 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.900035 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.915007 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.950383 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.951897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.951941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.951956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.951974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.951986 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:37Z","lastTransitionTime":"2026-03-18T10:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.967586 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.983953 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:37 crc kubenswrapper[4733]: I0318 10:14:37.997780 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:37Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.020671 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.037549 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.054996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.055046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.055058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.055077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.055089 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.056262 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.070465 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.083928 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.100494 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.125658 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.139779 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.158300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.158344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.158354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.158370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.158382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.173045 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:37Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI0318 10:14:37.435628 6605 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435688 6605 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435917 6605 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436100 6605 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436433 6605 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.436884 6605 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.438358 6605 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 10:14:37.438401 6605 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 10:14:37.438470 6605 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 10:14:37.438508 6605 factory.go:656] Stopping watch factory\\\\nI0318 10:14:37.438541 6605 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.195637 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.261246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.261346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.261371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.261412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.261438 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.364819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.364877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.364904 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.364931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.364947 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.467704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.467733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.467741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.467753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.467762 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.570121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.570153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.570162 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.570224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.570233 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.672377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.672414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.672424 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.672440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.672450 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.774207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.774273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.774286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.774305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.774320 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.839253 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/0.log" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.842507 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.842906 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.859160 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.871404 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.876090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.876154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.876172 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.876237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.876271 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.883535 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.895511 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.911481 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.928870 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.943462 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.961151 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.975531 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.978078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.978113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.978122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.978139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.978149 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:38Z","lastTransitionTime":"2026-03-18T10:14:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.989342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:38 crc kubenswrapper[4733]: I0318 10:14:38.998368 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:38Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.029452 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:37Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI0318 10:14:37.435628 6605 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435688 6605 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435917 6605 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436100 6605 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436433 6605 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.436884 6605 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.438358 6605 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 10:14:37.438401 6605 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 10:14:37.438470 6605 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 10:14:37.438508 6605 factory.go:656] Stopping watch factory\\\\nI0318 10:14:37.438541 6605 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.061752 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.072390 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.080384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.080427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.080440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.080461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.080474 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.086336 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.098277 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.111251 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.121083 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.175336 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.175382 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.175409 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:39 crc kubenswrapper[4733]: E0318 10:14:39.175457 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.175469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:39 crc kubenswrapper[4733]: E0318 10:14:39.175551 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:39 crc kubenswrapper[4733]: E0318 10:14:39.175605 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:39 crc kubenswrapper[4733]: E0318 10:14:39.175656 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.182293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.182339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.182353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.182373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.182387 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.285423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.285505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.285534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.285567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.285593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.388702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.388741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.388755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.388771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.388785 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.492882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.492954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.493008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.493034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.493053 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.596422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.596487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.596506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.596532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.596550 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.699754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.699805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.699824 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.699845 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.699860 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.803731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.803774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.803786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.803804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.803816 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.848131 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/1.log" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.848958 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/0.log" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.852110 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2" exitCode=1 Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.852169 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.852263 4733 scope.go:117] "RemoveContainer" containerID="f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.853038 4733 scope.go:117] "RemoveContainer" containerID="e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2" Mar 18 10:14:39 crc kubenswrapper[4733]: E0318 10:14:39.853319 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.894477 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.906830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.906885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.906902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.906926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.906943 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:39Z","lastTransitionTime":"2026-03-18T10:14:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.907494 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.922902 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.949584 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.976090 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:39 crc kubenswrapper[4733]: I0318 10:14:39.995730 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:39Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.009324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.009363 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.009374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.009388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.009399 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.020171 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.038304 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.057022 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.073951 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.094047 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.111904 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.112156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.112180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.112207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.112221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.112232 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.126526 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.142405 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.160790 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.177830 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.194202 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.217508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.217562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.217572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.217587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.217598 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.221996 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f090e00a8f1c87ebbc1c282f9e7528c15f4755dce93436309480932f74815e96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:37Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI0318 10:14:37.435628 6605 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435688 6605 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.435917 6605 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436100 6605 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 10:14:37.436433 6605 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.436884 6605 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 10:14:37.438358 6605 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 10:14:37.438401 6605 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 10:14:37.438470 6605 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 10:14:37.438508 6605 factory.go:656] Stopping watch factory\\\\nI0318 10:14:37.438541 6605 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.319798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.319864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.319877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.319893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.319903 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.422547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.422597 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.422610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.422629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.422644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.525813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.525894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.525908 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.525927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.525938 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.628664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.628724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.628735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.628751 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.628760 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.732237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.732309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.732321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.732338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.732347 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.835442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.835501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.835516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.835537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.835552 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.857012 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/1.log" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.861179 4733 scope.go:117] "RemoveContainer" containerID="e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2" Mar 18 10:14:40 crc kubenswrapper[4733]: E0318 10:14:40.861324 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.873538 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.884616 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.896574 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.914033 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.931808 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.938031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.938066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.938078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.938116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.938129 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:40Z","lastTransitionTime":"2026-03-18T10:14:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.944350 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.955994 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.965606 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.976871 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:40 crc kubenswrapper[4733]: I0318 10:14:40.989068 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:40Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.006010 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.022675 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.040896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.040944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.040958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.040976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.040988 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:41Z","lastTransitionTime":"2026-03-18T10:14:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.042634 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.055571 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.071028 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.083004 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.101599 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.111861 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.141578 4733 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.174496 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.174613 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.174494 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.174493 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.174703 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.174716 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.174839 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.174900 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.186687 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.197446 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.207910 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.220052 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.232602 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.246480 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.256809 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: E0318 10:14:41.266478 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.273766 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.288988 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.299077 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.308528 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.325694 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.337102 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.353965 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.362569 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.372607 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.382553 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:41 crc kubenswrapper[4733]: I0318 10:14:41.394033 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:41Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.249467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.249775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.249785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.249800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.249811 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:42Z","lastTransitionTime":"2026-03-18T10:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.263133 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:42Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.267301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.267368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.267390 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.267414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.267432 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:42Z","lastTransitionTime":"2026-03-18T10:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.284321 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:42Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.288262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.288296 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.288308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.288323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.288333 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:42Z","lastTransitionTime":"2026-03-18T10:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.304300 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:42Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.308946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.308995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.309009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.309029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.309043 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:42Z","lastTransitionTime":"2026-03-18T10:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.323546 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:42Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.327139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.327217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.327225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.327238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:42 crc kubenswrapper[4733]: I0318 10:14:42.327247 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:42Z","lastTransitionTime":"2026-03-18T10:14:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.337920 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:42Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:42 crc kubenswrapper[4733]: E0318 10:14:42.338064 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:43 crc kubenswrapper[4733]: I0318 10:14:43.175394 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:43 crc kubenswrapper[4733]: E0318 10:14:43.175538 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:43 crc kubenswrapper[4733]: I0318 10:14:43.175583 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:43 crc kubenswrapper[4733]: I0318 10:14:43.175590 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:43 crc kubenswrapper[4733]: E0318 10:14:43.175753 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:43 crc kubenswrapper[4733]: E0318 10:14:43.175902 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:43 crc kubenswrapper[4733]: I0318 10:14:43.175936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:43 crc kubenswrapper[4733]: E0318 10:14:43.176062 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:44 crc kubenswrapper[4733]: I0318 10:14:44.187866 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.173412 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.173611 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173654 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.173626508 +0000 UTC m=+156.665360833 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.173707 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173773 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173863 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.173834423 +0000 UTC m=+156.665568788 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.173901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173970 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173985 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.173982 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.173996 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174085 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.17407014 +0000 UTC m=+156.665804505 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174179 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174270 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.174253405 +0000 UTC m=+156.665987760 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174332 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174372 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.174360528 +0000 UTC m=+156.666094893 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.174460 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.174507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174584 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.174644 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174729 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.174772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174896 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.174984 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:45 crc kubenswrapper[4733]: I0318 10:14:45.275228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.275427 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.275455 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.275466 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:45 crc kubenswrapper[4733]: E0318 10:14:45.275525 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:15:17.275508981 +0000 UTC m=+156.767243306 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:14:46 crc kubenswrapper[4733]: E0318 10:14:46.268094 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:14:47 crc kubenswrapper[4733]: I0318 10:14:47.175301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:47 crc kubenswrapper[4733]: E0318 10:14:47.175464 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:47 crc kubenswrapper[4733]: I0318 10:14:47.175974 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:47 crc kubenswrapper[4733]: E0318 10:14:47.176081 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:47 crc kubenswrapper[4733]: I0318 10:14:47.176154 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:47 crc kubenswrapper[4733]: E0318 10:14:47.176286 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:47 crc kubenswrapper[4733]: I0318 10:14:47.176506 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:47 crc kubenswrapper[4733]: E0318 10:14:47.176769 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:49 crc kubenswrapper[4733]: I0318 10:14:49.175039 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:49 crc kubenswrapper[4733]: I0318 10:14:49.175117 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:49 crc kubenswrapper[4733]: I0318 10:14:49.175123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:49 crc kubenswrapper[4733]: I0318 10:14:49.175036 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:49 crc kubenswrapper[4733]: E0318 10:14:49.175309 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:49 crc kubenswrapper[4733]: E0318 10:14:49.176297 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:49 crc kubenswrapper[4733]: E0318 10:14:49.176544 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:49 crc kubenswrapper[4733]: E0318 10:14:49.176800 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.174686 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:51 crc kubenswrapper[4733]: E0318 10:14:51.174897 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.174919 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.174968 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:51 crc kubenswrapper[4733]: E0318 10:14:51.175096 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.174947 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:51 crc kubenswrapper[4733]: E0318 10:14:51.175312 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:51 crc kubenswrapper[4733]: E0318 10:14:51.175438 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.197478 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.213385 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.227804 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.244431 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: E0318 10:14:51.268847 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.277026 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.304752 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.324215 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.337375 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.355079 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.376204 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.386694 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.419941 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.436271 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.450031 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.462472 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.477077 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.498070 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.511794 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:51 crc kubenswrapper[4733]: I0318 10:14:51.531587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:51Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.497025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.497058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.497067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.497080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.497088 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:52Z","lastTransitionTime":"2026-03-18T10:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.515536 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:52Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.519519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.519554 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.519564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.519578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.519590 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:52Z","lastTransitionTime":"2026-03-18T10:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.536990 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:52Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.540853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.540945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.540970 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.541003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.541026 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:52Z","lastTransitionTime":"2026-03-18T10:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.561103 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:52Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.565998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.566066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.566091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.566122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.566147 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:52Z","lastTransitionTime":"2026-03-18T10:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.582121 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:52Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.587979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.588046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.588093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.588127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:14:52 crc kubenswrapper[4733]: I0318 10:14:52.588152 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:14:52Z","lastTransitionTime":"2026-03-18T10:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.603084 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:52Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:52 crc kubenswrapper[4733]: E0318 10:14:52.603208 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:14:53 crc kubenswrapper[4733]: I0318 10:14:53.175060 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:53 crc kubenswrapper[4733]: I0318 10:14:53.175109 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:53 crc kubenswrapper[4733]: I0318 10:14:53.175067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:53 crc kubenswrapper[4733]: I0318 10:14:53.175067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:53 crc kubenswrapper[4733]: E0318 10:14:53.175275 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:53 crc kubenswrapper[4733]: E0318 10:14:53.175392 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:53 crc kubenswrapper[4733]: E0318 10:14:53.175485 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:53 crc kubenswrapper[4733]: E0318 10:14:53.175864 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.175524 4733 scope.go:117] "RemoveContainer" containerID="e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.916831 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/1.log" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.919356 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6"} Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.919799 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.931877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.942016 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.950484 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.960741 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.976935 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.989474 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:54 crc kubenswrapper[4733]: I0318 10:14:54.999176 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:54Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.007466 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.016848 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.031318 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.040726 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.066143 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.074706 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.085244 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.094856 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.104815 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.115387 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.126915 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.138828 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.174869 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:55 crc kubenswrapper[4733]: E0318 10:14:55.174994 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.175162 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.175243 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:55 crc kubenswrapper[4733]: E0318 10:14:55.175253 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:55 crc kubenswrapper[4733]: E0318 10:14:55.175313 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.174882 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:55 crc kubenswrapper[4733]: E0318 10:14:55.175382 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.925084 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/2.log" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.926019 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/1.log" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.929163 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" exitCode=1 Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.929229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6"} Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.929330 4733 scope.go:117] "RemoveContainer" containerID="e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.929835 4733 scope.go:117] "RemoveContainer" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" Mar 18 10:14:55 crc kubenswrapper[4733]: E0318 10:14:55.929980 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.942780 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.956859 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.972846 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:55 crc kubenswrapper[4733]: I0318 10:14:55.987898 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:55Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.007421 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.025320 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.038855 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.061699 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e92f95b4e7499a2bf80c4498b1b592c2cb8a47a2602131b74b47ee685f9562f2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:38Z\\\",\\\"message\\\":\\\"xfvfl\\\\nI0318 10:14:38.896757 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896763 6729 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xfvfl in node crc\\\\nI0318 10:14:38.896768 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-xfvfl after 0 failed attempt(s)\\\\nI0318 10:14:38.896772 6729 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-xfvfl\\\\nI0318 10:14:38.896804 6729 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896809 6729 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896815 6729 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-g6j2q in node crc\\\\nI0318 10:14:38.896820 6729 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-g6j2q after 0 failed attempt(s)\\\\nI0318 10:14:38.896824 6729 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-g6j2q\\\\nI0318 10:14:38.896848 6729 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0318 10:14:38.896915 6729 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.086440 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.100389 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.122352 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.139781 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.161175 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.176234 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.192121 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.208509 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.224002 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.238700 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.255831 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: E0318 10:14:56.270720 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.935910 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/2.log" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.942835 4733 scope.go:117] "RemoveContainer" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" Mar 18 10:14:56 crc kubenswrapper[4733]: E0318 10:14:56.943417 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.966674 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:56 crc kubenswrapper[4733]: I0318 10:14:56.986285 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:56Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.004403 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.019374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.033793 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.051817 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.063491 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.078298 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.097706 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.113289 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.133381 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.147521 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.171967 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.176306 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.176473 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.176321 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:57 crc kubenswrapper[4733]: E0318 10:14:57.176566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:57 crc kubenswrapper[4733]: E0318 10:14:57.176707 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.176760 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:57 crc kubenswrapper[4733]: E0318 10:14:57.176886 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:57 crc kubenswrapper[4733]: E0318 10:14:57.177046 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.190214 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.218271 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.231840 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.248673 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.267768 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:57 crc kubenswrapper[4733]: I0318 10:14:57.284342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:14:57Z is after 2025-08-24T17:21:41Z" Mar 18 10:14:59 crc kubenswrapper[4733]: I0318 10:14:59.175473 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:14:59 crc kubenswrapper[4733]: I0318 10:14:59.175488 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:14:59 crc kubenswrapper[4733]: E0318 10:14:59.176235 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:14:59 crc kubenswrapper[4733]: I0318 10:14:59.175562 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:14:59 crc kubenswrapper[4733]: E0318 10:14:59.176343 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:14:59 crc kubenswrapper[4733]: I0318 10:14:59.175527 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:14:59 crc kubenswrapper[4733]: E0318 10:14:59.176442 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:14:59 crc kubenswrapper[4733]: E0318 10:14:59.176516 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.175477 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:01 crc kubenswrapper[4733]: E0318 10:15:01.175673 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.176305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.176370 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:01 crc kubenswrapper[4733]: E0318 10:15:01.176399 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:01 crc kubenswrapper[4733]: E0318 10:15:01.176525 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.176760 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:01 crc kubenswrapper[4733]: E0318 10:15:01.177062 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.188760 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.210958 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.222583 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.239280 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.253731 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: E0318 10:15:01.271328 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.275245 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.293909 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.304311 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.314873 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.325401 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.337759 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.348788 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.357859 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.367692 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.381964 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.398315 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.413659 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.427309 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:01 crc kubenswrapper[4733]: I0318 10:15:01.451100 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:01Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.655925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.656010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.656023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.656051 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.656069 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:02Z","lastTransitionTime":"2026-03-18T10:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.675082 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:02Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.680016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.680090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.680104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.680129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.680144 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:02Z","lastTransitionTime":"2026-03-18T10:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.697743 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:02Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.701690 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.701787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.701818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.701854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.701881 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:02Z","lastTransitionTime":"2026-03-18T10:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.718177 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:02Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.722282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.722318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.722331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.722355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.722369 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:02Z","lastTransitionTime":"2026-03-18T10:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.738335 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:02Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.743042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.743114 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.743136 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.743164 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:02 crc kubenswrapper[4733]: I0318 10:15:02.743227 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:02Z","lastTransitionTime":"2026-03-18T10:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.760068 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:02Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:02 crc kubenswrapper[4733]: E0318 10:15:02.760349 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:15:03 crc kubenswrapper[4733]: I0318 10:15:03.175447 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:03 crc kubenswrapper[4733]: I0318 10:15:03.175568 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:03 crc kubenswrapper[4733]: I0318 10:15:03.175601 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:03 crc kubenswrapper[4733]: E0318 10:15:03.175702 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:03 crc kubenswrapper[4733]: I0318 10:15:03.175763 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:03 crc kubenswrapper[4733]: E0318 10:15:03.175887 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:03 crc kubenswrapper[4733]: E0318 10:15:03.175968 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:03 crc kubenswrapper[4733]: E0318 10:15:03.176021 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:05 crc kubenswrapper[4733]: I0318 10:15:05.175065 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:05 crc kubenswrapper[4733]: I0318 10:15:05.175138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:05 crc kubenswrapper[4733]: I0318 10:15:05.175237 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:05 crc kubenswrapper[4733]: E0318 10:15:05.176555 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:05 crc kubenswrapper[4733]: E0318 10:15:05.176718 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:05 crc kubenswrapper[4733]: I0318 10:15:05.175272 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:05 crc kubenswrapper[4733]: E0318 10:15:05.176931 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:05 crc kubenswrapper[4733]: E0318 10:15:05.177139 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:06 crc kubenswrapper[4733]: E0318 10:15:06.272751 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:07 crc kubenswrapper[4733]: I0318 10:15:07.174814 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:07 crc kubenswrapper[4733]: I0318 10:15:07.174929 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:07 crc kubenswrapper[4733]: I0318 10:15:07.174929 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:07 crc kubenswrapper[4733]: E0318 10:15:07.175571 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:07 crc kubenswrapper[4733]: E0318 10:15:07.175556 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:07 crc kubenswrapper[4733]: I0318 10:15:07.175081 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:07 crc kubenswrapper[4733]: E0318 10:15:07.175863 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:07 crc kubenswrapper[4733]: E0318 10:15:07.176308 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:09 crc kubenswrapper[4733]: I0318 10:15:09.175045 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:09 crc kubenswrapper[4733]: E0318 10:15:09.175313 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:09 crc kubenswrapper[4733]: I0318 10:15:09.175539 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:09 crc kubenswrapper[4733]: I0318 10:15:09.175646 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:09 crc kubenswrapper[4733]: E0318 10:15:09.175781 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:09 crc kubenswrapper[4733]: I0318 10:15:09.175802 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:09 crc kubenswrapper[4733]: E0318 10:15:09.176056 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:09 crc kubenswrapper[4733]: E0318 10:15:09.176137 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.174629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.174741 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.175469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.175570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.175667 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.175689 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.175851 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.175908 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.176675 4733 scope.go:117] "RemoveContainer" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.176849 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.196469 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.210322 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.224138 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.237846 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.255401 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: E0318 10:15:11.274913 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.280116 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.294034 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.313272 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.322645 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.334663 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.347609 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.360403 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.373636 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.395366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.408778 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.425612 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.441157 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.455601 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:11 crc kubenswrapper[4733]: I0318 10:15:11.470455 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:11Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.866621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.866664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.866679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.866695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.866708 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:12Z","lastTransitionTime":"2026-03-18T10:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.882245 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:12Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.886998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.887042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.887057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.887077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.887095 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:12Z","lastTransitionTime":"2026-03-18T10:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.905114 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:12Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.909883 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.909961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.909984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.910017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.910041 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:12Z","lastTransitionTime":"2026-03-18T10:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.924365 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:12Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.929334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.929381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.929396 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.929417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.929431 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:12Z","lastTransitionTime":"2026-03-18T10:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.943848 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:12Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.948587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.948663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.948686 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.948717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:12 crc kubenswrapper[4733]: I0318 10:15:12.948740 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:12Z","lastTransitionTime":"2026-03-18T10:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.965637 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:12Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:12 crc kubenswrapper[4733]: E0318 10:15:12.965786 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.002824 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/0.log" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.002917 4733 generic.go:334] "Generic (PLEG): container finished" podID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" containerID="cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034" exitCode=1 Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.002982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerDied","Data":"cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034"} Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.003713 4733 scope.go:117] "RemoveContainer" containerID="cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.021370 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.043580 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.061916 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.089939 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.103582 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.118979 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.132314 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.149102 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.164970 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.174724 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.174776 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.174887 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:13 crc kubenswrapper[4733]: E0318 10:15:13.174898 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:13 crc kubenswrapper[4733]: E0318 10:15:13.175019 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:13 crc kubenswrapper[4733]: E0318 10:15:13.175149 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.175305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:13 crc kubenswrapper[4733]: E0318 10:15:13.175422 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.177522 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.193061 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.207034 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.220637 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.233457 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.249320 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.271238 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.288163 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.304871 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:13 crc kubenswrapper[4733]: I0318 10:15:13.317752 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:13Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.009004 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/0.log" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.009074 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerStarted","Data":"b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a"} Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.025017 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.044713 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.062612 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.079667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.102777 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.118482 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.138344 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.152052 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.170152 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.207340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.226816 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.246403 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.262363 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.281045 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.297700 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.316042 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.346897 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.363894 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:14 crc kubenswrapper[4733]: I0318 10:15:14.382400 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:14Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:15 crc kubenswrapper[4733]: I0318 10:15:15.174502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:15 crc kubenswrapper[4733]: I0318 10:15:15.174590 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:15 crc kubenswrapper[4733]: E0318 10:15:15.174650 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:15 crc kubenswrapper[4733]: I0318 10:15:15.174770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:15 crc kubenswrapper[4733]: E0318 10:15:15.174806 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:15 crc kubenswrapper[4733]: I0318 10:15:15.174837 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:15 crc kubenswrapper[4733]: E0318 10:15:15.175026 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:15 crc kubenswrapper[4733]: E0318 10:15:15.175155 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:16 crc kubenswrapper[4733]: E0318 10:15:16.276676 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.174646 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.174705 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.174721 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.174896 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.174929 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.175114 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.175325 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.175458 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.227466 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.227617 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.227732 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.227696183 +0000 UTC m=+220.719430558 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.227823 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.227831 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.227857 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.227886 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.227893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.227964 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.227939931 +0000 UTC m=+220.719674296 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.228007 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228029 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228097 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.228078335 +0000 UTC m=+220.719812700 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228149 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228228 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228255 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.22823218 +0000 UTC m=+220.719966535 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.228296 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.228277271 +0000 UTC m=+220.720011636 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: I0318 10:15:17.328836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.329030 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.329061 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.329080 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:15:17 crc kubenswrapper[4733]: E0318 10:15:17.329163 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:21.329141008 +0000 UTC m=+220.820875373 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:15:19 crc kubenswrapper[4733]: I0318 10:15:19.175410 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:19 crc kubenswrapper[4733]: I0318 10:15:19.175548 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:19 crc kubenswrapper[4733]: I0318 10:15:19.175626 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:19 crc kubenswrapper[4733]: I0318 10:15:19.175616 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:19 crc kubenswrapper[4733]: E0318 10:15:19.176916 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:19 crc kubenswrapper[4733]: E0318 10:15:19.177108 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:19 crc kubenswrapper[4733]: E0318 10:15:19.177286 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:19 crc kubenswrapper[4733]: E0318 10:15:19.176752 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.175210 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:21 crc kubenswrapper[4733]: E0318 10:15:21.175330 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.175395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.175449 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.175517 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:21 crc kubenswrapper[4733]: E0318 10:15:21.175472 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:21 crc kubenswrapper[4733]: E0318 10:15:21.175633 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:21 crc kubenswrapper[4733]: E0318 10:15:21.175685 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.196334 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.213792 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.248113 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.266872 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: E0318 10:15:21.277074 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.283111 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.296281 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.306438 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.315106 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.325597 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.341555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.356910 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.370137 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.378670 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.391111 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.404141 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.414756 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.436437 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.446937 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:21 crc kubenswrapper[4733]: I0318 10:15:21.460368 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:21Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.153160 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.153264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.153293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.153316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.153333 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:23Z","lastTransitionTime":"2026-03-18T10:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.175304 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.175422 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.175594 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.175775 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.175906 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.175981 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:23Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.176454 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.177001 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.177348 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.181646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.181707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.181729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.181752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.181770 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:23Z","lastTransitionTime":"2026-03-18T10:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.202327 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:23Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.207454 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.207507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.207528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.207553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.207574 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:23Z","lastTransitionTime":"2026-03-18T10:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.228676 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:23Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.233877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.234097 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.234275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.234436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.234580 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:23Z","lastTransitionTime":"2026-03-18T10:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.255477 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:23Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.260961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.261238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.261583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.261759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:23 crc kubenswrapper[4733]: I0318 10:15:23.261909 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:23Z","lastTransitionTime":"2026-03-18T10:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.280209 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:23Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:23 crc kubenswrapper[4733]: E0318 10:15:23.280328 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:15:24 crc kubenswrapper[4733]: I0318 10:15:24.176565 4733 scope.go:117] "RemoveContainer" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.052985 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/2.log" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.055619 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.056402 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.075617 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.090279 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.107332 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.128898 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.140827 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.152132 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.167954 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.174884 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.174935 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.175017 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:25 crc kubenswrapper[4733]: E0318 10:15:25.175098 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:25 crc kubenswrapper[4733]: E0318 10:15:25.175256 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:25 crc kubenswrapper[4733]: E0318 10:15:25.175323 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.175443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:25 crc kubenswrapper[4733]: E0318 10:15:25.175591 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.191821 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.208019 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.220073 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.242248 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.253882 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.266837 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.281384 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.297521 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.312325 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.325914 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.344009 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:25 crc kubenswrapper[4733]: I0318 10:15:25.357181 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:25Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.062932 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/3.log" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.064003 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/2.log" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.067848 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" exitCode=1 Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.067909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.067959 4733 scope.go:117] "RemoveContainer" containerID="b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.069094 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:15:26 crc kubenswrapper[4733]: E0318 10:15:26.069367 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.091787 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.107342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.128036 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.142145 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.165022 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.181902 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.193180 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.204797 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.216988 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.232064 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.243374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.253771 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: E0318 10:15:26.278044 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.282140 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a81a8bb9ca8ad4c87fd9b3cd1ae0f5c21d0e4b39a32bd67b6c63b41175d0a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:14:54Z\\\",\\\"message\\\":\\\"\\\\nI0318 10:14:54.959635 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0318 10:14:54.959527 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959642 6930 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI0318 10:14:54.959646 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-etcd/etcd-crc after 0 failed attempt(s)\\\\nI0318 10:14:54.959649 6930 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI0318 10:14:54.959466 6930 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0318 10:14:54.959658 6930 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0318 10:14:54.959662 6930 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0318 10:14:54.959665 6930 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF0318 10:14:54.959243 6930 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:25Z\\\",\\\"message\\\":\\\"_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"3ec9f67e-7758-4707-a6d0-2dc28f28ac37\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0318 10:15:25.172380 7260 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:15:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.305495 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.316483 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.328045 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.340599 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.360651 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:26 crc kubenswrapper[4733]: I0318 10:15:26.376266 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:26Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.074782 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/3.log" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.081146 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:15:27 crc kubenswrapper[4733]: E0318 10:15:27.081494 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.098277 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.118069 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.138310 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.171499 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.174579 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.174629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.174697 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:27 crc kubenswrapper[4733]: E0318 10:15:27.174823 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.174859 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:27 crc kubenswrapper[4733]: E0318 10:15:27.175028 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:27 crc kubenswrapper[4733]: E0318 10:15:27.175179 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:27 crc kubenswrapper[4733]: E0318 10:15:27.175384 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.195587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.213818 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.233915 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.247801 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.263368 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.294998 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:25Z\\\",\\\"message\\\":\\\"_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"3ec9f67e-7758-4707-a6d0-2dc28f28ac37\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0318 10:15:25.172380 7260 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:15:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.316307 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.333983 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.348247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.365559 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.391184 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.402829 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.426007 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.441308 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:27 crc kubenswrapper[4733]: I0318 10:15:27.459971 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:27Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:29 crc kubenswrapper[4733]: I0318 10:15:29.175246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:29 crc kubenswrapper[4733]: I0318 10:15:29.175320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:29 crc kubenswrapper[4733]: I0318 10:15:29.175331 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:29 crc kubenswrapper[4733]: E0318 10:15:29.175434 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:29 crc kubenswrapper[4733]: I0318 10:15:29.175522 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:29 crc kubenswrapper[4733]: E0318 10:15:29.175549 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:29 crc kubenswrapper[4733]: E0318 10:15:29.175832 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:29 crc kubenswrapper[4733]: E0318 10:15:29.176304 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.174756 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.174883 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:31 crc kubenswrapper[4733]: E0318 10:15:31.174958 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.174795 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.174888 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:31 crc kubenswrapper[4733]: E0318 10:15:31.175104 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:31 crc kubenswrapper[4733]: E0318 10:15:31.175291 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:31 crc kubenswrapper[4733]: E0318 10:15:31.175459 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.208027 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b23fcc-38c7-420b-ad9a-57d1c547c788\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb888f7a23904596729e28ec137231447f22565be42be8589f1481aa52efd9f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d0f02cb69f907a82795f47bfae39d1f750bb7bedeeb6d0802e84087dd7150df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cfda710da166c7b27fe6df3f38f5f969d0edea58503530ace9d35e3a7ec1420\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b454a77a46e10fcea3615e1f59d7849430a461ee7392b37fbbb6ec89e53eb432\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://448f3d210c3e435bb68acc8f81dd92e63739d073e0d3746be3985c3d3fe07556\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa633c3540b7c0b9ac0a8ccaca1fb619e1fffff85dd8626f4140c926862ba1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://980838b633a0f6d7144508883fbb8308ffbdba2b27ce7eecf1e046f28b30ec6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1b9a779bc130375be7c7d0792015ec5a467cfc85fd2bead18618dc5e292b2e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.220878 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"908bd772-fb33-4f68-8971-d1fef3118c82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3457636bb3e1cc25507158454524b9cee6812beb56c7b22fb86b9438b8082488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5452bdbefc77b8eab2e8b5f9f71d0edb627db0661edd4abd3989970363812fb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.235682 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a37904dce4f31563b6bf3db4a4e779fcaebf12e80cdabf402fb1fcf03320f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.250029 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.267340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t28sh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f82588a-9dbd-4c55-8cfc-f96e57fa58b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14c8bb1225c6c415d19ccaf11f0117aa22ccf43aa3b80472a8779ec5cea1aeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f534576a07c40f1e53709418bbc816e439ac9e036c97e780922c19e36c272642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5819ab8cd6ee5f116abf7861120afb8ac702ade05eb0efc4f73366042f2dd3bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50025671e1a19a96b715db63cbfa653f3cf078d3bfcfd94803bca2fac9636637\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bb1ae24448e6a65c00cce421e6bbc772e3cd734f8f35fb411ce8eac5d662256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaee328f999419b8d4f3a3298dbf38fd35bc92f76224d38c6769aa7426bb9a80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a0b2c5f56088e948c02d27d94da94aba67e2c6ffc58442adc30586a548271b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xwk4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t28sh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: E0318 10:15:31.281891 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.281667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f75e1c5-e0c5-43df-944f-77b734070793\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b4eaa631b67f13321cd60f9136da1832c5cd6e226609c01cabfa28410630a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpnv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2h7dp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.298897 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddb303e3-8922-4b43-9bba-2d3f0c30c6b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 10:13:41.916017 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 10:13:41.916132 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 10:13:41.917022 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1943543564/tls.crt::/tmp/serving-cert-1943543564/tls.key\\\\\\\"\\\\nI0318 10:13:42.070462 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 10:13:42.072416 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 10:13:42.072438 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 10:13:42.072464 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 10:13:42.072469 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 10:13:42.076902 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 10:13:42.076943 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076949 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 10:13:42.076959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 10:13:42.076962 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 10:13:42.076967 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 10:13:42.076974 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 10:13:42.077028 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 10:13:42.078631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:13:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.311455 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e797e62-fc82-47f7-8c8c-6c11d3463304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0cb2e53d9e61f6e93594f61ef9614e057a66575c32d18a010ab1ecfd3ac367f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fe29241779e03381bb946ac650ea8a793785c0c3ed67302dd89f1c5e0d93e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5b09c8d5c3c63eb7d9db92ce941aec0f0def87adbc1d46334ccc518a47c60f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b4b403598b0be68c5baba6e126ecad218005a9c2aeea9badf14dfc4859dce03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:12:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.324441 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.333987 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4s425" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3650177-e338-4eba-ab42-bc0cd14c9d65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x9zpb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4s425\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.347655 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a14e8a496af63cf1951ed21cfb3b13b1b516b00271dce19cdf858148beff398b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61bc78e89fc84025b585b2a421fa96e8da9f90840b8c78c0658f30d8738c64ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.364783 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d693a73-68c1-4595-bbcc-be97691b06fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb07463e9cec5d204a136bc3da2a197f348b611ad242f9652741da372ebc490f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e784035b634ef119368039982dbafab7f160c3864fe9ef9f5236d906de281b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vg7hc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-spfjj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.379362 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hsk58" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2c181c8-3361-40a2-afc5-a677e0ab4ecd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7ffcba189533d7ca155ab3284efac3d072ee3bc46d4b2a61247261bdaecb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-httph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hsk58\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.395018 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aaa6e82080eecc5cde4d763e00b69fb4234de74431affa584f0b900a811dd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.417217 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g6j2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:12Z\\\",\\\"message\\\":\\\"2026-03-18T10:14:27+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c\\\\n2026-03-18T10:14:27+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_92f48003-aa44-42d0-a76f-02756a51562c to /host/opt/cni/bin/\\\\n2026-03-18T10:14:27Z [verbose] multus-daemon started\\\\n2026-03-18T10:14:27Z [verbose] Readiness Indicator file check\\\\n2026-03-18T10:15:12Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:15:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ph8vv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g6j2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.438872 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"353ee984-b20f-41fa-978a-0167c20ede36\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:13:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:12:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4287a7d43815108131e4b725925805740a64682bc2a9c96ff054f65517e501f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7c73fedb720681572ba31d10e49b7fc28537f98b4afb32bee611e6265eafaff\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T10:13:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 10:12:43.210581 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 10:12:43.213660 1 observer_polling.go:159] Starting file observer\\\\nI0318 10:12:43.251533 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 10:12:43.256315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 10:13:13.491530 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:13:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b54b5cea02ea38b404d6b5730afbab0f729978207023e1dfa7cc49ea9736795\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b29e8c4af41ef6391d7ea79821c7caa64424b8113473541a96ae936db10015\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:12:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:12:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.453715 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.468432 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xfvfl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb58b528-9013-4fab-9747-60bb6ff1bc1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc72346f1bb873e40a1063486ebd2adfd16e3958e17730370c00cb3b775a982c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zg7jp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xfvfl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:31 crc kubenswrapper[4733]: I0318 10:15:31.493754 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73327417-4d3b-45f1-b3b6-575fdeeaa31a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T10:14:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T10:15:25Z\\\",\\\"message\\\":\\\"_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"3ec9f67e-7758-4707-a6d0-2dc28f28ac37\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0318 10:15:25.172380 7260 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T10:15:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T10:14:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T10:14:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T10:14:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zqxdr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T10:14:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7pxwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:31Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.175239 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.175349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.175404 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.175477 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.175425 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.175648 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.175723 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.175819 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.619641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.619681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.619692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.619709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.619720 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:33Z","lastTransitionTime":"2026-03-18T10:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.632147 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.635976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.636049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.636076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.636106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.636130 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:33Z","lastTransitionTime":"2026-03-18T10:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.657969 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.662875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.662930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.662945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.662967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.662984 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:33Z","lastTransitionTime":"2026-03-18T10:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.679404 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.683374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.683429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.683447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.683469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.683484 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:33Z","lastTransitionTime":"2026-03-18T10:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.702659 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.707046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.707090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.707103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.707121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:33 crc kubenswrapper[4733]: I0318 10:15:33.707133 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:33Z","lastTransitionTime":"2026-03-18T10:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.724887 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T10:15:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a826494-c246-4717-869b-fd136e2b8410\\\",\\\"systemUUID\\\":\\\"fe704b25-4cdf-410a-9afb-ebc7963f4bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T10:15:33Z is after 2025-08-24T17:21:41Z" Mar 18 10:15:33 crc kubenswrapper[4733]: E0318 10:15:33.724996 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 10:15:35 crc kubenswrapper[4733]: I0318 10:15:35.175318 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:35 crc kubenswrapper[4733]: I0318 10:15:35.175384 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:35 crc kubenswrapper[4733]: E0318 10:15:35.175542 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:35 crc kubenswrapper[4733]: I0318 10:15:35.175584 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:35 crc kubenswrapper[4733]: E0318 10:15:35.175718 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:35 crc kubenswrapper[4733]: I0318 10:15:35.175320 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:35 crc kubenswrapper[4733]: E0318 10:15:35.175864 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:35 crc kubenswrapper[4733]: E0318 10:15:35.176153 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:36 crc kubenswrapper[4733]: E0318 10:15:36.283961 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:37 crc kubenswrapper[4733]: I0318 10:15:37.175167 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:37 crc kubenswrapper[4733]: I0318 10:15:37.175323 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:37 crc kubenswrapper[4733]: E0318 10:15:37.175360 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:37 crc kubenswrapper[4733]: I0318 10:15:37.175477 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:37 crc kubenswrapper[4733]: E0318 10:15:37.175688 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:37 crc kubenswrapper[4733]: I0318 10:15:37.175809 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:37 crc kubenswrapper[4733]: E0318 10:15:37.175921 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:37 crc kubenswrapper[4733]: E0318 10:15:37.176004 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:38 crc kubenswrapper[4733]: I0318 10:15:38.176453 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:15:38 crc kubenswrapper[4733]: E0318 10:15:38.176737 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:15:39 crc kubenswrapper[4733]: I0318 10:15:39.175554 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:39 crc kubenswrapper[4733]: I0318 10:15:39.175593 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:39 crc kubenswrapper[4733]: I0318 10:15:39.175717 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:39 crc kubenswrapper[4733]: E0318 10:15:39.175750 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:39 crc kubenswrapper[4733]: E0318 10:15:39.175894 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:39 crc kubenswrapper[4733]: I0318 10:15:39.175987 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:39 crc kubenswrapper[4733]: E0318 10:15:39.176032 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:39 crc kubenswrapper[4733]: E0318 10:15:39.176239 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.174977 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.175063 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.175146 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:41 crc kubenswrapper[4733]: E0318 10:15:41.175382 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.175468 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:41 crc kubenswrapper[4733]: E0318 10:15:41.175614 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:41 crc kubenswrapper[4733]: E0318 10:15:41.176519 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:41 crc kubenswrapper[4733]: E0318 10:15:41.176678 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.213554 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-g6j2q" podStartSLOduration=123.213535372 podStartE2EDuration="2m3.213535372s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.210940249 +0000 UTC m=+180.702674574" watchObservedRunningTime="2026-03-18 10:15:41.213535372 +0000 UTC m=+180.705269697" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.243133 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hsk58" podStartSLOduration=123.243110148 podStartE2EDuration="2m3.243110148s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.243036746 +0000 UTC m=+180.734771081" watchObservedRunningTime="2026-03-18 10:15:41.243110148 +0000 UTC m=+180.734844473" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.243369 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-spfjj" podStartSLOduration=123.243361866 podStartE2EDuration="2m3.243361866s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.225452793 +0000 UTC m=+180.717187208" watchObservedRunningTime="2026-03-18 10:15:41.243361866 +0000 UTC m=+180.735096191" Mar 18 10:15:41 crc kubenswrapper[4733]: E0318 10:15:41.287610 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.306833 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=57.306819136 podStartE2EDuration="57.306819136s" podCreationTimestamp="2026-03-18 10:14:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.306056802 +0000 UTC m=+180.797791127" watchObservedRunningTime="2026-03-18 10:15:41.306819136 +0000 UTC m=+180.798553461" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.341231 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xfvfl" podStartSLOduration=123.341144854 podStartE2EDuration="2m3.341144854s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.340521164 +0000 UTC m=+180.832255529" watchObservedRunningTime="2026-03-18 10:15:41.341144854 +0000 UTC m=+180.832879229" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.378746 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-t28sh" podStartSLOduration=123.378719836 podStartE2EDuration="2m3.378719836s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.376854096 +0000 UTC m=+180.868588461" watchObservedRunningTime="2026-03-18 10:15:41.378719836 +0000 UTC m=+180.870454161" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.421402 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.421377801 podStartE2EDuration="1m19.421377801s" podCreationTimestamp="2026-03-18 10:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.42106005 +0000 UTC m=+180.912794405" watchObservedRunningTime="2026-03-18 10:15:41.421377801 +0000 UTC m=+180.913112126" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.421906 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podStartSLOduration=123.421888687 podStartE2EDuration="2m3.421888687s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.394536332 +0000 UTC m=+180.886270697" watchObservedRunningTime="2026-03-18 10:15:41.421888687 +0000 UTC m=+180.913623012" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.432930 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=88.432894119 podStartE2EDuration="1m28.432894119s" podCreationTimestamp="2026-03-18 10:14:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.43135764 +0000 UTC m=+180.923091965" watchObservedRunningTime="2026-03-18 10:15:41.432894119 +0000 UTC m=+180.924628484" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.498930 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.498909051 podStartE2EDuration="1m19.498909051s" podCreationTimestamp="2026-03-18 10:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.49856666 +0000 UTC m=+180.990300995" watchObservedRunningTime="2026-03-18 10:15:41.498909051 +0000 UTC m=+180.990643366" Mar 18 10:15:41 crc kubenswrapper[4733]: I0318 10:15:41.510940 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=66.510917455 podStartE2EDuration="1m6.510917455s" podCreationTimestamp="2026-03-18 10:14:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:41.510609745 +0000 UTC m=+181.002344070" watchObservedRunningTime="2026-03-18 10:15:41.510917455 +0000 UTC m=+181.002651790" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.174821 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.174898 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.174956 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:43 crc kubenswrapper[4733]: E0318 10:15:43.175055 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.175072 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:43 crc kubenswrapper[4733]: E0318 10:15:43.175205 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:43 crc kubenswrapper[4733]: E0318 10:15:43.175301 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:43 crc kubenswrapper[4733]: E0318 10:15:43.175357 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.792038 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.792099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.792118 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.792145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.792165 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T10:15:43Z","lastTransitionTime":"2026-03-18T10:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.865536 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7"] Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.866394 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.869382 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.869706 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.873326 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 10:15:43 crc kubenswrapper[4733]: I0318 10:15:43.873395 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.046171 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.046322 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.046394 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.046477 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.046601 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.147854 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.147911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.147955 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.147999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.148045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.148136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.148184 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.149830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.158059 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.169580 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ed168c9-7562-4ab5-8cf4-5edfa11200e7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nsd7\" (UID: \"1ed168c9-7562-4ab5-8cf4-5edfa11200e7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.192270 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.282207 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 18 10:15:44 crc kubenswrapper[4733]: I0318 10:15:44.292065 4733 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.143757 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" event={"ID":"1ed168c9-7562-4ab5-8cf4-5edfa11200e7","Type":"ContainerStarted","Data":"af3b957f0d8068e24e6b30b30a154402543b1c9b867238d2a05edf65767bee90"} Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.143824 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" event={"ID":"1ed168c9-7562-4ab5-8cf4-5edfa11200e7","Type":"ContainerStarted","Data":"4bfa48ba8fbdea9f5c215496267acbf7b6181ea4a50d5f4a1e09661215475b0c"} Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.175015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.175029 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:45 crc kubenswrapper[4733]: E0318 10:15:45.175389 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.175162 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:45 crc kubenswrapper[4733]: E0318 10:15:45.175741 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:45 crc kubenswrapper[4733]: E0318 10:15:45.175619 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:45 crc kubenswrapper[4733]: I0318 10:15:45.175063 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:45 crc kubenswrapper[4733]: E0318 10:15:45.176001 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:46 crc kubenswrapper[4733]: E0318 10:15:46.288776 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:47 crc kubenswrapper[4733]: I0318 10:15:47.174679 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:47 crc kubenswrapper[4733]: I0318 10:15:47.174687 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:47 crc kubenswrapper[4733]: I0318 10:15:47.174770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:47 crc kubenswrapper[4733]: I0318 10:15:47.174908 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:47 crc kubenswrapper[4733]: E0318 10:15:47.175120 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:47 crc kubenswrapper[4733]: E0318 10:15:47.175270 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:47 crc kubenswrapper[4733]: E0318 10:15:47.175354 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:47 crc kubenswrapper[4733]: E0318 10:15:47.175466 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:49 crc kubenswrapper[4733]: I0318 10:15:49.174926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:49 crc kubenswrapper[4733]: I0318 10:15:49.174958 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:49 crc kubenswrapper[4733]: E0318 10:15:49.175051 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:49 crc kubenswrapper[4733]: I0318 10:15:49.174927 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:49 crc kubenswrapper[4733]: I0318 10:15:49.175140 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:49 crc kubenswrapper[4733]: E0318 10:15:49.176266 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:49 crc kubenswrapper[4733]: I0318 10:15:49.179503 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:15:49 crc kubenswrapper[4733]: E0318 10:15:49.179866 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:49 crc kubenswrapper[4733]: E0318 10:15:49.180248 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:15:49 crc kubenswrapper[4733]: E0318 10:15:49.180484 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:51 crc kubenswrapper[4733]: I0318 10:15:51.174874 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:51 crc kubenswrapper[4733]: I0318 10:15:51.176162 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:51 crc kubenswrapper[4733]: E0318 10:15:51.176356 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:51 crc kubenswrapper[4733]: I0318 10:15:51.176414 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:51 crc kubenswrapper[4733]: E0318 10:15:51.176951 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:51 crc kubenswrapper[4733]: I0318 10:15:51.176446 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:51 crc kubenswrapper[4733]: E0318 10:15:51.177100 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:51 crc kubenswrapper[4733]: E0318 10:15:51.177417 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:51 crc kubenswrapper[4733]: E0318 10:15:51.289406 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:53 crc kubenswrapper[4733]: I0318 10:15:53.174536 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:53 crc kubenswrapper[4733]: I0318 10:15:53.174608 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:53 crc kubenswrapper[4733]: I0318 10:15:53.174562 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:53 crc kubenswrapper[4733]: E0318 10:15:53.174762 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:53 crc kubenswrapper[4733]: I0318 10:15:53.174790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:53 crc kubenswrapper[4733]: E0318 10:15:53.175389 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:53 crc kubenswrapper[4733]: E0318 10:15:53.175552 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:53 crc kubenswrapper[4733]: E0318 10:15:53.175755 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:55 crc kubenswrapper[4733]: I0318 10:15:55.174923 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:55 crc kubenswrapper[4733]: I0318 10:15:55.174945 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:55 crc kubenswrapper[4733]: I0318 10:15:55.175052 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:55 crc kubenswrapper[4733]: E0318 10:15:55.175246 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:55 crc kubenswrapper[4733]: E0318 10:15:55.175461 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:55 crc kubenswrapper[4733]: E0318 10:15:55.175623 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:55 crc kubenswrapper[4733]: I0318 10:15:55.176237 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:55 crc kubenswrapper[4733]: E0318 10:15:55.176378 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:56 crc kubenswrapper[4733]: E0318 10:15:56.291263 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:15:57 crc kubenswrapper[4733]: I0318 10:15:57.175411 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:57 crc kubenswrapper[4733]: E0318 10:15:57.175570 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:57 crc kubenswrapper[4733]: I0318 10:15:57.175773 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:57 crc kubenswrapper[4733]: E0318 10:15:57.175830 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:57 crc kubenswrapper[4733]: I0318 10:15:57.175954 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:57 crc kubenswrapper[4733]: E0318 10:15:57.176014 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:57 crc kubenswrapper[4733]: I0318 10:15:57.176178 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:57 crc kubenswrapper[4733]: E0318 10:15:57.176390 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.175419 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.175514 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.175538 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:15:59 crc kubenswrapper[4733]: E0318 10:15:59.175632 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:15:59 crc kubenswrapper[4733]: E0318 10:15:59.175833 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:15:59 crc kubenswrapper[4733]: E0318 10:15:59.175880 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.176046 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:15:59 crc kubenswrapper[4733]: E0318 10:15:59.176114 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.228597 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/1.log" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.229075 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/0.log" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.229120 4733 generic.go:334] "Generic (PLEG): container finished" podID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" containerID="b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a" exitCode=1 Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.229150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerDied","Data":"b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a"} Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.229205 4733 scope.go:117] "RemoveContainer" containerID="cf9836f3455051ee686f0ec11ceb1c60cff06c95a16bf2fcff6c4c3ed600b034" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.229852 4733 scope.go:117] "RemoveContainer" containerID="b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a" Mar 18 10:15:59 crc kubenswrapper[4733]: E0318 10:15:59.230127 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-g6j2q_openshift-multus(cc85b0d4-15a5-4894-9f07-9aaeb28f63fa)\"" pod="openshift-multus/multus-g6j2q" podUID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" Mar 18 10:15:59 crc kubenswrapper[4733]: I0318 10:15:59.250963 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nsd7" podStartSLOduration=141.250937344 podStartE2EDuration="2m21.250937344s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:15:45.162241315 +0000 UTC m=+184.653975640" watchObservedRunningTime="2026-03-18 10:15:59.250937344 +0000 UTC m=+198.742671709" Mar 18 10:16:00 crc kubenswrapper[4733]: I0318 10:16:00.234515 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/1.log" Mar 18 10:16:01 crc kubenswrapper[4733]: I0318 10:16:01.175184 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:01 crc kubenswrapper[4733]: I0318 10:16:01.175257 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:01 crc kubenswrapper[4733]: I0318 10:16:01.175243 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:01 crc kubenswrapper[4733]: E0318 10:16:01.176099 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:01 crc kubenswrapper[4733]: I0318 10:16:01.176232 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:01 crc kubenswrapper[4733]: E0318 10:16:01.176264 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:01 crc kubenswrapper[4733]: E0318 10:16:01.176383 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:01 crc kubenswrapper[4733]: E0318 10:16:01.176531 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:01 crc kubenswrapper[4733]: E0318 10:16:01.291797 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:16:03 crc kubenswrapper[4733]: I0318 10:16:03.175082 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:03 crc kubenswrapper[4733]: E0318 10:16:03.175332 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:03 crc kubenswrapper[4733]: I0318 10:16:03.175533 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:03 crc kubenswrapper[4733]: I0318 10:16:03.175709 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:03 crc kubenswrapper[4733]: E0318 10:16:03.175874 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:03 crc kubenswrapper[4733]: I0318 10:16:03.175912 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:03 crc kubenswrapper[4733]: E0318 10:16:03.176078 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:03 crc kubenswrapper[4733]: E0318 10:16:03.176299 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:04 crc kubenswrapper[4733]: I0318 10:16:04.175900 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:16:04 crc kubenswrapper[4733]: E0318 10:16:04.176076 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7pxwd_openshift-ovn-kubernetes(73327417-4d3b-45f1-b3b6-575fdeeaa31a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" Mar 18 10:16:05 crc kubenswrapper[4733]: I0318 10:16:05.175405 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:05 crc kubenswrapper[4733]: I0318 10:16:05.175483 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:05 crc kubenswrapper[4733]: I0318 10:16:05.175492 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:05 crc kubenswrapper[4733]: E0318 10:16:05.175608 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:05 crc kubenswrapper[4733]: E0318 10:16:05.175757 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:05 crc kubenswrapper[4733]: I0318 10:16:05.175839 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:05 crc kubenswrapper[4733]: E0318 10:16:05.175912 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:05 crc kubenswrapper[4733]: E0318 10:16:05.176065 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:06 crc kubenswrapper[4733]: E0318 10:16:06.293315 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:16:07 crc kubenswrapper[4733]: I0318 10:16:07.175248 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:07 crc kubenswrapper[4733]: I0318 10:16:07.175340 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:07 crc kubenswrapper[4733]: I0318 10:16:07.175406 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:07 crc kubenswrapper[4733]: I0318 10:16:07.175266 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:07 crc kubenswrapper[4733]: E0318 10:16:07.175492 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:07 crc kubenswrapper[4733]: E0318 10:16:07.175585 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:07 crc kubenswrapper[4733]: E0318 10:16:07.175664 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:07 crc kubenswrapper[4733]: E0318 10:16:07.175808 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:09 crc kubenswrapper[4733]: I0318 10:16:09.175522 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:09 crc kubenswrapper[4733]: I0318 10:16:09.175635 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:09 crc kubenswrapper[4733]: I0318 10:16:09.175679 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:09 crc kubenswrapper[4733]: E0318 10:16:09.175713 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:09 crc kubenswrapper[4733]: I0318 10:16:09.175768 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:09 crc kubenswrapper[4733]: E0318 10:16:09.175919 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:09 crc kubenswrapper[4733]: E0318 10:16:09.176030 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:09 crc kubenswrapper[4733]: E0318 10:16:09.176089 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:11 crc kubenswrapper[4733]: I0318 10:16:11.174871 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:11 crc kubenswrapper[4733]: E0318 10:16:11.175866 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:11 crc kubenswrapper[4733]: I0318 10:16:11.175982 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:11 crc kubenswrapper[4733]: I0318 10:16:11.176039 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:11 crc kubenswrapper[4733]: I0318 10:16:11.175976 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:11 crc kubenswrapper[4733]: E0318 10:16:11.176127 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:11 crc kubenswrapper[4733]: E0318 10:16:11.176153 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:11 crc kubenswrapper[4733]: E0318 10:16:11.176232 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:11 crc kubenswrapper[4733]: E0318 10:16:11.293894 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:16:13 crc kubenswrapper[4733]: I0318 10:16:13.175226 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:13 crc kubenswrapper[4733]: I0318 10:16:13.175282 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:13 crc kubenswrapper[4733]: I0318 10:16:13.175261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:13 crc kubenswrapper[4733]: I0318 10:16:13.175247 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:13 crc kubenswrapper[4733]: E0318 10:16:13.175423 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:13 crc kubenswrapper[4733]: E0318 10:16:13.175748 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:13 crc kubenswrapper[4733]: E0318 10:16:13.175989 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:13 crc kubenswrapper[4733]: E0318 10:16:13.176225 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:14 crc kubenswrapper[4733]: I0318 10:16:14.175306 4733 scope.go:117] "RemoveContainer" containerID="b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.175349 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.175387 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:15 crc kubenswrapper[4733]: E0318 10:16:15.175562 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.175597 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.175610 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:15 crc kubenswrapper[4733]: E0318 10:16:15.175763 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:15 crc kubenswrapper[4733]: E0318 10:16:15.175873 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:15 crc kubenswrapper[4733]: E0318 10:16:15.175979 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.284102 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/1.log" Mar 18 10:16:15 crc kubenswrapper[4733]: I0318 10:16:15.284216 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerStarted","Data":"e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d"} Mar 18 10:16:16 crc kubenswrapper[4733]: E0318 10:16:16.295464 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:16:17 crc kubenswrapper[4733]: I0318 10:16:17.174753 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:17 crc kubenswrapper[4733]: I0318 10:16:17.174803 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:17 crc kubenswrapper[4733]: E0318 10:16:17.174929 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:17 crc kubenswrapper[4733]: I0318 10:16:17.174961 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:17 crc kubenswrapper[4733]: E0318 10:16:17.175376 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:17 crc kubenswrapper[4733]: E0318 10:16:17.175466 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:17 crc kubenswrapper[4733]: I0318 10:16:17.175671 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:17 crc kubenswrapper[4733]: I0318 10:16:17.176015 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:16:17 crc kubenswrapper[4733]: E0318 10:16:17.176271 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.048638 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4s425"] Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.049155 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:18 crc kubenswrapper[4733]: E0318 10:16:18.049335 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.300682 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/3.log" Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.303802 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerStarted","Data":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.304266 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:16:18 crc kubenswrapper[4733]: I0318 10:16:18.341110 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podStartSLOduration=160.34109251 podStartE2EDuration="2m40.34109251s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:18.340310945 +0000 UTC m=+217.832045280" watchObservedRunningTime="2026-03-18 10:16:18.34109251 +0000 UTC m=+217.832826845" Mar 18 10:16:19 crc kubenswrapper[4733]: I0318 10:16:19.175248 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:19 crc kubenswrapper[4733]: I0318 10:16:19.175362 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:19 crc kubenswrapper[4733]: E0318 10:16:19.175436 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:19 crc kubenswrapper[4733]: E0318 10:16:19.175498 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:19 crc kubenswrapper[4733]: I0318 10:16:19.175559 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:19 crc kubenswrapper[4733]: E0318 10:16:19.175743 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:20 crc kubenswrapper[4733]: I0318 10:16:20.174650 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:20 crc kubenswrapper[4733]: E0318 10:16:20.174862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4s425" podUID="b3650177-e338-4eba-ab42-bc0cd14c9d65" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.174684 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.174685 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.176610 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.176646 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.176831 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.176964 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.275507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.275828 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.275776184 +0000 UTC m=+342.767510519 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.275954 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.276014 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.276074 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.276125 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276234 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276315 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276333 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.276320332 +0000 UTC m=+342.768054667 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276522 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.276459676 +0000 UTC m=+342.768194041 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276709 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276764 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276785 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276907 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.276872159 +0000 UTC m=+342.768606624 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.276999 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.277107 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs podName:b3650177-e338-4eba-ab42-bc0cd14c9d65 nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.277079076 +0000 UTC m=+342.768813441 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs") pod "network-metrics-daemon-4s425" (UID: "b3650177-e338-4eba-ab42-bc0cd14c9d65") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: I0318 10:16:21.377040 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.377372 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.377437 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.377459 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:16:21 crc kubenswrapper[4733]: E0318 10:16:21.377561 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 10:18:23.377533809 +0000 UTC m=+342.869268174 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.367944 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.368089 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.368777 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.374017 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.374064 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.374305 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 10:16:22 crc kubenswrapper[4733]: I0318 10:16:22.374456 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 10:16:23 crc kubenswrapper[4733]: I0318 10:16:23.175350 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:16:23 crc kubenswrapper[4733]: I0318 10:16:23.178301 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 10:16:23 crc kubenswrapper[4733]: I0318 10:16:23.178762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.589309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.645146 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.645903 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.648092 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-848w7"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.648977 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.652040 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nbftd"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.653459 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.653724 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.654858 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.655313 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.655436 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.655722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.661617 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-gxcb2"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662653 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662826 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662928 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662677 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662731 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.663285 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662802 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.663455 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.662846 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.666236 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.666746 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.666938 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.667173 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.678009 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.678469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.679075 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.680032 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.680922 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.681382 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.683531 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xh9n5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.683860 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.684395 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zztn5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.684899 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.686500 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.687305 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.687736 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.688345 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.691269 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-xl5d7"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.691988 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.692413 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.693521 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.693735 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.694110 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.701082 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.702477 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.702763 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.702937 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.703363 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.702830 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720014 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720287 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmgk\" (UniqueName: \"kubernetes.io/projected/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-kube-api-access-kgmgk\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43cea3fb-14f9-4993-a8a9-4618680e8286-serving-cert\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720547 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-images\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720653 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzj8w\" (UniqueName: \"kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720802 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.720914 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721011 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-config\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2wcj\" (UniqueName: \"kubernetes.io/projected/0c02459c-3d75-4363-a010-3e9639bb9b4e-kube-api-access-b2wcj\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721233 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w48br\" (UniqueName: \"kubernetes.io/projected/43cea3fb-14f9-4993-a8a9-4618680e8286-kube-api-access-w48br\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721399 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/43cea3fb-14f9-4993-a8a9-4618680e8286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721530 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.721995 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c02459c-3d75-4363-a010-3e9639bb9b4e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.751828 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.754613 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.754977 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.771569 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.772220 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.772301 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.773639 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.774970 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786258 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786373 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786405 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786511 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786523 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786661 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786673 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786785 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786840 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.786785 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790263 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790345 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790542 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790803 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.790945 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791058 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791117 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791226 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791314 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791419 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791489 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791553 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791641 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791671 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791769 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791804 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791890 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.791941 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.792019 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.792121 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.792169 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.792274 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.792336 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.812089 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813087 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813444 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813519 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813616 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813681 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813760 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813843 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813869 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813885 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.813934 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.814484 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.815308 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.815366 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.815566 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.815788 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.815861 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816006 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816162 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816290 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816449 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816625 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816770 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.816799 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.817762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.817801 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.817907 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.817927 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.818050 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.818218 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.818349 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.818569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.818812 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822835 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822880 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msfql\" (UniqueName: \"kubernetes.io/projected/61e27ee7-5eb0-4cc7-a696-85ddd192b171-kube-api-access-msfql\") pod \"downloads-7954f5f757-gxcb2\" (UID: \"61e27ee7-5eb0-4cc7-a696-85ddd192b171\") " pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822905 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-config\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822925 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822943 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-auth-proxy-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822966 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b6a9601-6689-435b-aca1-256a0c3c07fb-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.822987 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823009 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823030 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-metrics-certs\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823048 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c94c\" (UniqueName: \"kubernetes.io/projected/c0da800f-a7ca-4d0e-89bb-96673854969e-kube-api-access-5c94c\") pod \"migrator-59844c95c7-6572z\" (UID: \"c0da800f-a7ca-4d0e-89bb-96673854969e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z955v\" (UniqueName: \"kubernetes.io/projected/9c5f567e-b38f-44a0-b1fd-1a96857e811f-kube-api-access-z955v\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823085 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/158a5836-f175-4da3-b22d-6a3130a89d30-serving-cert\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823106 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c02459c-3d75-4363-a010-3e9639bb9b4e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823164 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-stats-auth\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823202 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-service-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2cdd\" (UniqueName: \"kubernetes.io/projected/99efba52-bc27-49d7-8efb-154b6e3787a9-kube-api-access-f2cdd\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823258 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2539fca8-3dde-43ed-815c-e837f37dfdd5-machine-approver-tls\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823278 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.823302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824102 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h5xdn"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824232 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwlw\" (UniqueName: \"kubernetes.io/projected/158a5836-f175-4da3-b22d-6a3130a89d30-kube-api-access-ngwlw\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/99efba52-bc27-49d7-8efb-154b6e3787a9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824296 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-serving-cert\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824356 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824377 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824402 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmgk\" (UniqueName: \"kubernetes.io/projected/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-kube-api-access-kgmgk\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43cea3fb-14f9-4993-a8a9-4618680e8286-serving-cert\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824447 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-images\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824471 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824492 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-client\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824511 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-default-certificate\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6a9601-6689-435b-aca1-256a0c3c07fb-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824557 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzj8w\" (UniqueName: \"kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824604 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdltm\" (UniqueName: \"kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824629 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddl6d\" (UniqueName: \"kubernetes.io/projected/2539fca8-3dde-43ed-815c-e837f37dfdd5-kube-api-access-ddl6d\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824650 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9jhd\" (UniqueName: \"kubernetes.io/projected/d9572819-3894-4603-bd2b-7c9465bb0067-kube-api-access-l9jhd\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824672 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c5f567e-b38f-44a0-b1fd-1a96857e811f-service-ca-bundle\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824693 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824712 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-service-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2wcj\" (UniqueName: \"kubernetes.io/projected/0c02459c-3d75-4363-a010-3e9639bb9b4e-kube-api-access-b2wcj\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824755 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-config\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824775 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w48br\" (UniqueName: \"kubernetes.io/projected/43cea3fb-14f9-4993-a8a9-4618680e8286-kube-api-access-w48br\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824795 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-config\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824531 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.825235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/43cea3fb-14f9-4993-a8a9-4618680e8286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.826130 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.831082 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.831611 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-config\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.832120 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c02459c-3d75-4363-a010-3e9639bb9b4e-images\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.832390 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.824814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/43cea3fb-14f9-4993-a8a9-4618680e8286-available-featuregates\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.832701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.832749 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s2kq\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-kube-api-access-6s2kq\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.833703 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.834473 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.834868 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.835230 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.835570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.835593 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vsnq2"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.835705 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.836134 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.836646 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.837049 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.837297 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.837548 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.837798 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.840607 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.842364 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.843660 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.847072 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.848460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.848847 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43cea3fb-14f9-4993-a8a9-4618680e8286-serving-cert\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.848996 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c02459c-3d75-4363-a010-3e9639bb9b4e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.849030 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.850508 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.851110 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.851673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.852010 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.853633 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.853663 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.853864 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.853932 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.854073 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.854118 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.854148 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.868724 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.874806 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lptjf"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.875494 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.876917 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h8kqf"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.877078 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.878289 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.878856 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.879373 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.884888 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.886651 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.888367 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.888789 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.889637 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563816-4582s"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.892851 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.895489 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t95b6"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.896513 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.896842 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-n6hmz"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.900386 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.901178 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.906146 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzj8w\" (UniqueName: \"kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w\") pod \"route-controller-manager-6576b87f9c-m5k95\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.911726 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.912403 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.912894 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.912987 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.913777 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.916957 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-848w7"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.920701 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xvnwv"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.921479 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.921587 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.924607 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gxcb2"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.926418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmgk\" (UniqueName: \"kubernetes.io/projected/ef9e43d5-8b80-4934-82b6-c8ee0591e1bf-kube-api-access-kgmgk\") pod \"openshift-apiserver-operator-796bbdcf4f-7z2vw\" (UID: \"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.926506 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.929445 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.932316 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.933995 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddl6d\" (UniqueName: \"kubernetes.io/projected/2539fca8-3dde-43ed-815c-e837f37dfdd5-kube-api-access-ddl6d\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934048 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9jhd\" (UniqueName: \"kubernetes.io/projected/d9572819-3894-4603-bd2b-7c9465bb0067-kube-api-access-l9jhd\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c5f567e-b38f-44a0-b1fd-1a96857e811f-service-ca-bundle\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934103 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-service-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934137 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934174 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-config\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934244 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s2kq\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-kube-api-access-6s2kq\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934281 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934315 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msfql\" (UniqueName: \"kubernetes.io/projected/61e27ee7-5eb0-4cc7-a696-85ddd192b171-kube-api-access-msfql\") pod \"downloads-7954f5f757-gxcb2\" (UID: \"61e27ee7-5eb0-4cc7-a696-85ddd192b171\") " pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-config\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934362 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d48vf\" (UniqueName: \"kubernetes.io/projected/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-kube-api-access-d48vf\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-auth-proxy-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934406 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934428 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-encryption-config\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934451 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lrt\" (UniqueName: \"kubernetes.io/projected/352d0ed5-c43b-431f-bd66-1749ab30d013-kube-api-access-l4lrt\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934473 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b6a9601-6689-435b-aca1-256a0c3c07fb-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934500 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934522 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934553 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-serving-cert\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934576 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfbq5\" (UniqueName: \"kubernetes.io/projected/9571ba80-f267-46ed-8d16-e44531cb0ce8-kube-api-access-wfbq5\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-policies\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934622 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934644 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-dir\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934677 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-metrics-certs\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934720 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c94c\" (UniqueName: \"kubernetes.io/projected/c0da800f-a7ca-4d0e-89bb-96673854969e-kube-api-access-5c94c\") pod \"migrator-59844c95c7-6572z\" (UID: \"c0da800f-a7ca-4d0e-89bb-96673854969e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/158a5836-f175-4da3-b22d-6a3130a89d30-serving-cert\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z955v\" (UniqueName: \"kubernetes.io/projected/9c5f567e-b38f-44a0-b1fd-1a96857e811f-kube-api-access-z955v\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934790 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934812 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934833 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqrn\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934855 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-client\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934878 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-service-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934902 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-stats-auth\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2cdd\" (UniqueName: \"kubernetes.io/projected/99efba52-bc27-49d7-8efb-154b6e3787a9-kube-api-access-f2cdd\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.934975 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2539fca8-3dde-43ed-815c-e837f37dfdd5-machine-approver-tls\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935037 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935056 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935090 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-srv-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935159 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwlw\" (UniqueName: \"kubernetes.io/projected/158a5836-f175-4da3-b22d-6a3130a89d30-kube-api-access-ngwlw\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935222 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/99efba52-bc27-49d7-8efb-154b6e3787a9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935245 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935266 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-serving-cert\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935288 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935393 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-proxy-tls\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935416 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-client\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935499 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6a9601-6689-435b-aca1-256a0c3c07fb-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935543 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-default-certificate\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.935581 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdltm\" (UniqueName: \"kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.937381 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-config\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.937501 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.938033 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.938068 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-service-ca\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.938382 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-auth-proxy-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.938460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.938550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c5f567e-b38f-44a0-b1fd-1a96857e811f-service-ca-bundle\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.936853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-service-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.939054 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.939374 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.939814 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6a9601-6689-435b-aca1-256a0c3c07fb-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.939957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.941259 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.941319 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/158a5836-f175-4da3-b22d-6a3130a89d30-config\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.941777 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2539fca8-3dde-43ed-815c-e837f37dfdd5-config\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.942951 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2539fca8-3dde-43ed-815c-e837f37dfdd5-machine-approver-tls\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.942978 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b6a9601-6689-435b-aca1-256a0c3c07fb-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.943278 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-metrics-certs\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: E0318 10:16:24.943447 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.443426925 +0000 UTC m=+224.935161360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.943525 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.944349 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-default-certificate\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.944413 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/99efba52-bc27-49d7-8efb-154b6e3787a9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.944923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-serving-cert\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.944941 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zztn5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.945040 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/158a5836-f175-4da3-b22d-6a3130a89d30-serving-cert\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.945150 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.946591 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nbftd"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.947708 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.947931 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.949740 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xh9n5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.951946 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.953147 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.954552 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lptjf"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.955535 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563816-4582s"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.957311 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9572819-3894-4603-bd2b-7c9465bb0067-etcd-client\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.958071 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.959838 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9c5f567e-b38f-44a0-b1fd-1a96857e811f-stats-auth\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.959956 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4b5s"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.961178 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2wcj\" (UniqueName: \"kubernetes.io/projected/0c02459c-3d75-4363-a010-3e9639bb9b4e-kube-api-access-b2wcj\") pod \"machine-api-operator-5694c8668f-nbftd\" (UID: \"0c02459c-3d75-4363-a010-3e9639bb9b4e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.961673 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.961828 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.964138 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hvmrz"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.964852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.965639 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.967229 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.967432 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.976480 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.978592 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h5xdn"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.979665 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t95b6"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.980595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w48br\" (UniqueName: \"kubernetes.io/projected/43cea3fb-14f9-4993-a8a9-4618680e8286-kube-api-access-w48br\") pod \"openshift-config-operator-7777fb866f-848w7\" (UID: \"43cea3fb-14f9-4993-a8a9-4618680e8286\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.980784 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.982093 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.983382 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.984745 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.985840 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.987336 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.988732 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.989137 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d"] Mar 18 10:16:24 crc kubenswrapper[4733]: I0318 10:16:24.993231 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.000762 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hvmrz"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.006310 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vsnq2"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.006373 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.008116 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.008151 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.012002 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.014477 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h8kqf"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.017414 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xvnwv"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.017454 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-n6hmz"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.020602 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4b5s"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.021503 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pvlch"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.022562 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.022870 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fnzxw"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.023280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.024316 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pvlch"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.028430 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036452 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036727 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036773 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036800 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036816 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036833 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqrn\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036890 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036915 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzlbf\" (UniqueName: \"kubernetes.io/projected/aa4b5542-dc36-4c93-88e5-a080729b94ae-kube-api-access-dzlbf\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-node-pullsecrets\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.036970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.037034 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.536996328 +0000 UTC m=+225.028730653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037105 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b405f127-b181-49a1-8205-aafd58d1fa7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037180 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037507 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037584 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px4sd\" (UniqueName: \"kubernetes.io/projected/d5979b3d-b4b4-4081-b486-4fcf91f6367c-kube-api-access-px4sd\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037623 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84ddb369-1909-4d63-a0c0-b250490992c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037651 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-cabundle\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037714 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad648fa7-2560-4aa0-8634-05bcbc48916f-config\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037750 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037767 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-registration-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037783 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87157be2-0fc3-4120-b9b6-d4494ace940a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037796 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-socket-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037820 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7522\" (UniqueName: \"kubernetes.io/projected/ad648fa7-2560-4aa0-8634-05bcbc48916f-kube-api-access-n7522\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037837 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037855 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6nw8\" (UniqueName: \"kubernetes.io/projected/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-kube-api-access-h6nw8\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037872 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037889 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-srv-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjbj\" (UniqueName: \"kubernetes.io/projected/9b0edb65-3bcf-484f-9707-d8124df1ec88-kube-api-access-ncjbj\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037935 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.037981 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4clx\" (UniqueName: \"kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038001 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-mountpoint-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f3d99e-72fa-4c62-8190-059d7a0effd1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038259 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-metrics-tls\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038278 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad648fa7-2560-4aa0-8634-05bcbc48916f-serving-cert\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-profile-collector-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038308 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038350 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038366 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4810c2fd-346b-44a0-b985-46d302060373-serving-cert\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038426 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-config\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038463 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-encryption-config\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lrt\" (UniqueName: \"kubernetes.io/projected/352d0ed5-c43b-431f-bd66-1749ab30d013-kube-api-access-l4lrt\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038509 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038524 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-plugins-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038565 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mscv\" (UniqueName: \"kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038582 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-policies\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.038630 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ffe185-3f09-44d0-a173-f95bb53c419e-config\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040311 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-policies\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040388 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2kgf\" (UniqueName: \"kubernetes.io/projected/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-kube-api-access-r2kgf\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040422 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfv2\" (UniqueName: \"kubernetes.io/projected/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-kube-api-access-ppfv2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040465 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-trusted-ca\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040499 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040549 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f3d99e-72fa-4c62-8190-059d7a0effd1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040631 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040678 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040721 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040862 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngmrr\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-kube-api-access-ngmrr\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040905 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97ffe185-3f09-44d0-a173-f95bb53c419e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.040998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b405f127-b181-49a1-8205-aafd58d1fa7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041108 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-client\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041609 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041646 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-csi-data-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041676 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-images\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.041697 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.541678708 +0000 UTC m=+225.033413163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041791 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-srv-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.041892 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042052 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042070 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-key\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042116 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fblf6\" (UniqueName: \"kubernetes.io/projected/4810c2fd-346b-44a0-b985-46d302060373-kube-api-access-fblf6\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042135 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042586 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042888 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.042915 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.043546 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.043574 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.043595 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-proxy-tls\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044143 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kwjh\" (UniqueName: \"kubernetes.io/projected/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-kube-api-access-6kwjh\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044179 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scsd7\" (UniqueName: \"kubernetes.io/projected/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-kube-api-access-scsd7\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044257 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grpxc\" (UniqueName: \"kubernetes.io/projected/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-kube-api-access-grpxc\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044285 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll8g5\" (UniqueName: \"kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044338 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044552 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.044684 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046657 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4fz\" (UniqueName: \"kubernetes.io/projected/ed943a82-ef39-4ebc-9d76-09bb69f3b800-kube-api-access-km4fz\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046682 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046684 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046728 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zf8\" (UniqueName: \"kubernetes.io/projected/10f3d99e-72fa-4c62-8190-059d7a0effd1-kube-api-access-t4zf8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046750 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit-dir\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046800 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ffe185-3f09-44d0-a173-f95bb53c419e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046820 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046869 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046907 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046973 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn4vp\" (UniqueName: \"kubernetes.io/projected/3a0400a1-7e6b-4335-8819-586d7a460e3d-kube-api-access-vn4vp\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.046998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.047108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9l2x\" (UniqueName: \"kubernetes.io/projected/87157be2-0fc3-4120-b9b6-d4494ace940a-kube-api-access-j9l2x\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.048729 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.049145 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-encryption-config\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.049294 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-srv-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.049443 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.050135 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.052945 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-etcd-client\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053170 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9571ba80-f267-46ed-8d16-e44531cb0ce8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053441 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-proxy-tls\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053496 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053522 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053549 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053574 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-webhook-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053615 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053647 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2whnv\" (UniqueName: \"kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv\") pod \"auto-csr-approver-29563816-4582s\" (UID: \"71a70c3c-d483-43f4-9f54-10978c7f8cc8\") " pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053673 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-apiservice-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053695 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d48vf\" (UniqueName: \"kubernetes.io/projected/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-kube-api-access-d48vf\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053716 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw7q6\" (UniqueName: \"kubernetes.io/projected/4edea753-21f5-44fd-b183-daf03845dcd8-kube-api-access-gw7q6\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053748 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3a0400a1-7e6b-4335-8819-586d7a460e3d-tmpfs\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053769 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053786 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2qxh\" (UniqueName: \"kubernetes.io/projected/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-kube-api-access-c2qxh\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.053832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-serving-cert\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054233 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lfsh\" (UniqueName: \"kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054321 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfbq5\" (UniqueName: \"kubernetes.io/projected/9571ba80-f267-46ed-8d16-e44531cb0ce8-kube-api-access-wfbq5\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054425 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-dir\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.054680 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/352d0ed5-c43b-431f-bd66-1749ab30d013-audit-dir\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.058490 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-proxy-tls\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.062560 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/352d0ed5-c43b-431f-bd66-1749ab30d013-serving-cert\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.069250 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.095754 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.108868 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.130556 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.149163 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155103 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155320 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155346 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-node-pullsecrets\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155398 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzlbf\" (UniqueName: \"kubernetes.io/projected/aa4b5542-dc36-4c93-88e5-a080729b94ae-kube-api-access-dzlbf\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b405f127-b181-49a1-8205-aafd58d1fa7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155464 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px4sd\" (UniqueName: \"kubernetes.io/projected/d5979b3d-b4b4-4081-b486-4fcf91f6367c-kube-api-access-px4sd\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155504 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84ddb369-1909-4d63-a0c0-b250490992c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155522 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-cabundle\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155544 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad648fa7-2560-4aa0-8634-05bcbc48916f-config\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155562 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155578 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87157be2-0fc3-4120-b9b6-d4494ace940a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-socket-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155607 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-registration-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155624 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7522\" (UniqueName: \"kubernetes.io/projected/ad648fa7-2560-4aa0-8634-05bcbc48916f-kube-api-access-n7522\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155642 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155659 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6nw8\" (UniqueName: \"kubernetes.io/projected/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-kube-api-access-h6nw8\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155673 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155698 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-srv-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjbj\" (UniqueName: \"kubernetes.io/projected/9b0edb65-3bcf-484f-9707-d8124df1ec88-kube-api-access-ncjbj\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155747 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4clx\" (UniqueName: \"kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155770 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-mountpoint-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155784 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-metrics-tls\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155798 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad648fa7-2560-4aa0-8634-05bcbc48916f-serving-cert\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155815 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155846 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f3d99e-72fa-4c62-8190-059d7a0effd1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-profile-collector-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155880 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155906 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4810c2fd-346b-44a0-b985-46d302060373-serving-cert\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155921 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155935 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-config\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155956 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155971 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-plugins-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.155987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156002 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ffe185-3f09-44d0-a173-f95bb53c419e-config\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156016 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mscv\" (UniqueName: \"kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156034 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156050 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156066 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2kgf\" (UniqueName: \"kubernetes.io/projected/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-kube-api-access-r2kgf\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156082 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfv2\" (UniqueName: \"kubernetes.io/projected/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-kube-api-access-ppfv2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156096 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156114 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f3d99e-72fa-4c62-8190-059d7a0effd1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-trusted-ca\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156143 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97ffe185-3f09-44d0-a173-f95bb53c419e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156233 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156248 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156265 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngmrr\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-kube-api-access-ngmrr\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b405f127-b181-49a1-8205-aafd58d1fa7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156322 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-csi-data-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-images\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156395 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-key\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.156421 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.656406238 +0000 UTC m=+225.148140563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156461 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fblf6\" (UniqueName: \"kubernetes.io/projected/4810c2fd-346b-44a0-b985-46d302060373-kube-api-access-fblf6\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156478 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156512 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156572 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scsd7\" (UniqueName: \"kubernetes.io/projected/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-kube-api-access-scsd7\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kwjh\" (UniqueName: \"kubernetes.io/projected/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-kube-api-access-6kwjh\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156610 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grpxc\" (UniqueName: \"kubernetes.io/projected/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-kube-api-access-grpxc\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156628 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll8g5\" (UniqueName: \"kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156809 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156860 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4fz\" (UniqueName: \"kubernetes.io/projected/ed943a82-ef39-4ebc-9d76-09bb69f3b800-kube-api-access-km4fz\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156891 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156910 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zf8\" (UniqueName: \"kubernetes.io/projected/10f3d99e-72fa-4c62-8190-059d7a0effd1-kube-api-access-t4zf8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156931 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit-dir\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ffe185-3f09-44d0-a173-f95bb53c419e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156969 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.156986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157008 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-node-pullsecrets\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157376 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn4vp\" (UniqueName: \"kubernetes.io/projected/3a0400a1-7e6b-4335-8819-586d7a460e3d-kube-api-access-vn4vp\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157407 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157428 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9l2x\" (UniqueName: \"kubernetes.io/projected/87157be2-0fc3-4120-b9b6-d4494ace940a-kube-api-access-j9l2x\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-proxy-tls\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157517 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b405f127-b181-49a1-8205-aafd58d1fa7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157525 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157564 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-webhook-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157598 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157629 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2whnv\" (UniqueName: \"kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv\") pod \"auto-csr-approver-29563816-4582s\" (UID: \"71a70c3c-d483-43f4-9f54-10978c7f8cc8\") " pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-apiservice-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157675 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw7q6\" (UniqueName: \"kubernetes.io/projected/4edea753-21f5-44fd-b183-daf03845dcd8-kube-api-access-gw7q6\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157702 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3a0400a1-7e6b-4335-8819-586d7a460e3d-tmpfs\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157722 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157738 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157756 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2qxh\" (UniqueName: \"kubernetes.io/projected/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-kube-api-access-c2qxh\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157775 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157805 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lfsh\" (UniqueName: \"kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157850 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.157968 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-registration-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.158351 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-csi-data-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.158327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-mountpoint-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.158488 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-plugins-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.158530 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit-dir\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.159479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.159969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-key\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.159984 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.160163 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.160563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed943a82-ef39-4ebc-9d76-09bb69f3b800-socket-dir\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.161911 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-signing-cabundle\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.162586 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.162795 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.165385 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-profile-collector-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.166301 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.166495 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-images\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.167947 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.169286 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-proxy-tls\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.169379 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.169391 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3a0400a1-7e6b-4335-8819-586d7a460e3d-tmpfs\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.169654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.170134 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4edea753-21f5-44fd-b183-daf03845dcd8-srv-cert\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.170577 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.170679 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.172039 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.172232 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.188335 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.193685 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.208569 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.228593 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.229010 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.248381 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.264766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.265131 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.765117175 +0000 UTC m=+225.256851500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.268050 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.277748 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.278948 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.288502 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.295608 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.307785 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.328827 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.341775 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.348303 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.366357 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.367340 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.867322115 +0000 UTC m=+225.359056440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.368887 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.384658 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.390073 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.390506 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nbftd"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.401149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-metrics-tls\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:25 crc kubenswrapper[4733]: W0318 10:16:25.401807 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25331c44_b639_46f7_8a7f_6f62f8779e2b.slice/crio-0a4e2b2140bacea055efd9eb333f7f7f1da7235e623090af40eaf58bc070ecb2 WatchSource:0}: Error finding container 0a4e2b2140bacea055efd9eb333f7f7f1da7235e623090af40eaf58bc070ecb2: Status 404 returned error can't find the container with id 0a4e2b2140bacea055efd9eb333f7f7f1da7235e623090af40eaf58bc070ecb2 Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.402179 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" event={"ID":"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf","Type":"ContainerStarted","Data":"7d91e90576d6f69a752503d1db70731b645844079844bd8d008241f1b17a35ca"} Mar 18 10:16:25 crc kubenswrapper[4733]: W0318 10:16:25.404911 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c02459c_3d75_4363_a010_3e9639bb9b4e.slice/crio-1c6c59013156a64bebe9e58fcc19d04eef4644b14406fc14905ab52fb61b3f46 WatchSource:0}: Error finding container 1c6c59013156a64bebe9e58fcc19d04eef4644b14406fc14905ab52fb61b3f46: Status 404 returned error can't find the container with id 1c6c59013156a64bebe9e58fcc19d04eef4644b14406fc14905ab52fb61b3f46 Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.409125 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.431556 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.448979 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.452828 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b405f127-b181-49a1-8205-aafd58d1fa7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.468366 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-848w7"] Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.468578 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.469107 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.469724 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:25.96970877 +0000 UTC m=+225.461443095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.472450 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-webhook-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.474277 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a0400a1-7e6b-4335-8819-586d7a460e3d-apiservice-cert\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:25 crc kubenswrapper[4733]: W0318 10:16:25.475589 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43cea3fb_14f9_4993_a8a9_4618680e8286.slice/crio-266a4c82e657da88a4fe5fb06d7ebc997772f1e97435dac09b83d427850e9c87 WatchSource:0}: Error finding container 266a4c82e657da88a4fe5fb06d7ebc997772f1e97435dac09b83d427850e9c87: Status 404 returned error can't find the container with id 266a4c82e657da88a4fe5fb06d7ebc997772f1e97435dac09b83d427850e9c87 Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.488514 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.509241 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.528543 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.540158 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f3d99e-72fa-4c62-8190-059d7a0effd1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.548125 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.567784 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.570699 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.571026 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.0710055 +0000 UTC m=+225.562739825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.571609 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.571974 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.07195774 +0000 UTC m=+225.563692155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.573446 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f3d99e-72fa-4c62-8190-059d7a0effd1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.608399 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.636275 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.639322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-trusted-ca\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.648463 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.662015 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4810c2fd-346b-44a0-b985-46d302060373-config\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.669503 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.672362 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.673296 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.173282512 +0000 UTC m=+225.665016837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.688595 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.707910 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.724136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4810c2fd-346b-44a0-b985-46d302060373-serving-cert\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.728983 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.732706 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87157be2-0fc3-4120-b9b6-d4494ace940a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.748689 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.770289 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.775020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.775454 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.275435649 +0000 UTC m=+225.767170054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.782960 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97ffe185-3f09-44d0-a173-f95bb53c419e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.788724 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.808261 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.829021 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.831370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97ffe185-3f09-44d0-a173-f95bb53c419e-config\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.849359 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.869471 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.876773 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.877103 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.37706724 +0000 UTC m=+225.868801595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.877515 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.878100 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.378086793 +0000 UTC m=+225.869821118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.888501 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.906497 4733 request.go:700] Waited for 1.005027989s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.908908 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.928816 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.941781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad648fa7-2560-4aa0-8634-05bcbc48916f-serving-cert\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.948864 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.950207 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad648fa7-2560-4aa0-8634-05bcbc48916f-config\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.969477 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.979314 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.979470 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.479438165 +0000 UTC m=+225.971172500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.980564 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:25 crc kubenswrapper[4733]: E0318 10:16:25.981033 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.481021426 +0000 UTC m=+225.972755761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:25 crc kubenswrapper[4733]: I0318 10:16:25.991609 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.016960 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.028613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.028997 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.041444 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.050274 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.061472 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.068756 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.080861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.083774 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.084002 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.583970338 +0000 UTC m=+226.075704683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.084677 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.085167 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.585147986 +0000 UTC m=+226.076882331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.089016 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.093124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.110220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.123394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.143267 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.149157 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.153934 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158312 4733 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158343 4733 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158381 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca podName:486eda8c-6e6f-4761-b28c-8aeb72fcfcc1 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658363548 +0000 UTC m=+226.150097873 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-n6hmz" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158425 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert podName:9b0edb65-3bcf-484f-9707-d8124df1ec88 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658404099 +0000 UTC m=+226.150138434 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-kd6gw" (UID: "9b0edb65-3bcf-484f-9707-d8124df1ec88") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158424 4733 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158450 4733 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158470 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658459191 +0000 UTC m=+226.150193526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158487 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume podName:aa4b5542-dc36-4c93-88e5-a080729b94ae nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658477891 +0000 UTC m=+226.150212216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume") pod "dns-default-hvmrz" (UID: "aa4b5542-dc36-4c93-88e5-a080729b94ae") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158488 4733 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158501 4733 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158516 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token podName:a39a28f7-1fd2-44f7-8b49-05a0faf1e000 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658510922 +0000 UTC m=+226.150245247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token") pod "machine-config-server-fnzxw" (UID: "a39a28f7-1fd2-44f7-8b49-05a0faf1e000") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158318 4733 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158530 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs podName:a39a28f7-1fd2-44f7-8b49-05a0faf1e000 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658522113 +0000 UTC m=+226.150256458 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs") pod "machine-config-server-fnzxw" (UID: "a39a28f7-1fd2-44f7-8b49-05a0faf1e000") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158549 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658541093 +0000 UTC m=+226.150275438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158595 4733 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158624 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658617016 +0000 UTC m=+226.150351351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158644 4733 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158671 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert podName:d5979b3d-b4b4-4081-b486-4fcf91f6367c nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658662387 +0000 UTC m=+226.150396722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert") pod "ingress-canary-pvlch" (UID: "d5979b3d-b4b4-4081-b486-4fcf91f6367c") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158714 4733 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158741 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies podName:486eda8c-6e6f-4761-b28c-8aeb72fcfcc1 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658733919 +0000 UTC m=+226.150468254 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies") pod "oauth-openshift-558db77b4-n6hmz" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158760 4733 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158844 4733 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158874 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658866774 +0000 UTC m=+226.150601099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158919 4733 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158943 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert podName:84ddb369-1909-4d63-a0c0-b250490992c0 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.658937066 +0000 UTC m=+226.150671391 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert") pod "kube-controller-manager-operator-78b949d7b-bzhq6" (UID: "84ddb369-1909-4d63-a0c0-b250490992c0") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.158976 4733 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.159219 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.659179714 +0000 UTC m=+226.150914029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.159242 4733 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.159265 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config podName:84ddb369-1909-4d63-a0c0-b250490992c0 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.659258576 +0000 UTC m=+226.150992901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config") pod "kube-controller-manager-operator-78b949d7b-bzhq6" (UID: "84ddb369-1909-4d63-a0c0-b250490992c0") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160397 4733 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160404 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660367842 +0000 UTC m=+226.152102177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160483 4733 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160541 4733 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160498 4733 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160500 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session podName:486eda8c-6e6f-4761-b28c-8aeb72fcfcc1 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660481025 +0000 UTC m=+226.152215440 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session") pod "oauth-openshift-558db77b4-n6hmz" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160614 4733 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160641 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle podName:486eda8c-6e6f-4761-b28c-8aeb72fcfcc1 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.66061327 +0000 UTC m=+226.152347595 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-n6hmz" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160516 4733 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160657 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660652001 +0000 UTC m=+226.152386326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160706 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig podName:486eda8c-6e6f-4761-b28c-8aeb72fcfcc1 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660693112 +0000 UTC m=+226.152427547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-n6hmz" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160724 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls podName:aa4b5542-dc36-4c93-88e5-a080729b94ae nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660713853 +0000 UTC m=+226.152448278 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls") pod "dns-default-hvmrz" (UID: "aa4b5542-dc36-4c93-88e5-a080729b94ae") : failed to sync secret cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.160757 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca podName:56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6 nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.660748944 +0000 UTC m=+226.152483379 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca") pod "apiserver-76f77b778f-xvnwv" (UID: "56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6") : failed to sync configmap cache: timed out waiting for the condition Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.168875 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.185223 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.185849 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.685831346 +0000 UTC m=+226.177565671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.186024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.186374 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.686363433 +0000 UTC m=+226.178097758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.188777 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.207602 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.229235 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.256638 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.269436 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.287472 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.287692 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.787666874 +0000 UTC m=+226.279401199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.288097 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.289326 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.789279675 +0000 UTC m=+226.281014190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.289690 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.308990 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.329162 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.349046 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.368551 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.389145 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.389499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.389713 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.889683797 +0000 UTC m=+226.381418122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.390673 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.390974 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.890964438 +0000 UTC m=+226.382698753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.408407 4733 generic.go:334] "Generic (PLEG): container finished" podID="43cea3fb-14f9-4993-a8a9-4618680e8286" containerID="01b8744b846f9fb6ffe485cf34336b1d737fb2ba6bebf606b4c7d34295512d87" exitCode=0 Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.408488 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" event={"ID":"43cea3fb-14f9-4993-a8a9-4618680e8286","Type":"ContainerDied","Data":"01b8744b846f9fb6ffe485cf34336b1d737fb2ba6bebf606b4c7d34295512d87"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.408524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" event={"ID":"43cea3fb-14f9-4993-a8a9-4618680e8286","Type":"ContainerStarted","Data":"266a4c82e657da88a4fe5fb06d7ebc997772f1e97435dac09b83d427850e9c87"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.408691 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.410690 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" event={"ID":"0c02459c-3d75-4363-a010-3e9639bb9b4e","Type":"ContainerStarted","Data":"f90ad87e25013bf2f6f8581f6f43f0b35b6c4a22d87ae5d7ec104a7eda47afaa"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.410734 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" event={"ID":"0c02459c-3d75-4363-a010-3e9639bb9b4e","Type":"ContainerStarted","Data":"6418ec699be9d8b004cc9e2a35a84ccc8b26c389cdb4bbec509b2660f63b7fd4"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.410780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" event={"ID":"0c02459c-3d75-4363-a010-3e9639bb9b4e","Type":"ContainerStarted","Data":"1c6c59013156a64bebe9e58fcc19d04eef4644b14406fc14905ab52fb61b3f46"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.412749 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" event={"ID":"25331c44-b639-46f7-8a7f-6f62f8779e2b","Type":"ContainerStarted","Data":"3c09df7a275938153d455f147ffe12eff185edea72f1d8646898b9ba5ba684d5"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.412782 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" event={"ID":"25331c44-b639-46f7-8a7f-6f62f8779e2b","Type":"ContainerStarted","Data":"0a4e2b2140bacea055efd9eb333f7f7f1da7235e623090af40eaf58bc070ecb2"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.412968 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.414461 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" event={"ID":"ef9e43d5-8b80-4934-82b6-c8ee0591e1bf","Type":"ContainerStarted","Data":"ee668f2791d6ce76a1d528f6019a92f746e2d687878c1f727c7cad536298f5e3"} Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.428724 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.449376 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.469134 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.489453 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.492080 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.495070 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:26.995034607 +0000 UTC m=+226.486768932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.509698 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.530621 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.548251 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.574716 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.588339 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.595134 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.595491 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.09548036 +0000 UTC m=+226.587214685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.622040 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdltm\" (UniqueName: \"kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm\") pod \"console-f9d7485db-8v244\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.646767 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddl6d\" (UniqueName: \"kubernetes.io/projected/2539fca8-3dde-43ed-815c-e837f37dfdd5-kube-api-access-ddl6d\") pod \"machine-approver-56656f9798-9dd56\" (UID: \"2539fca8-3dde-43ed-815c-e837f37dfdd5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.662809 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msfql\" (UniqueName: \"kubernetes.io/projected/61e27ee7-5eb0-4cc7-a696-85ddd192b171-kube-api-access-msfql\") pod \"downloads-7954f5f757-gxcb2\" (UID: \"61e27ee7-5eb0-4cc7-a696-85ddd192b171\") " pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.689816 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9jhd\" (UniqueName: \"kubernetes.io/projected/d9572819-3894-4603-bd2b-7c9465bb0067-kube-api-access-l9jhd\") pod \"etcd-operator-b45778765-zztn5\" (UID: \"d9572819-3894-4603-bd2b-7c9465bb0067\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.691623 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.698058 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.698817 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.698991 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699122 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699556 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699675 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699787 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.699883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700162 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700340 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700684 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.700903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701445 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701573 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701712 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.701851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.702235 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.202215964 +0000 UTC m=+226.693950299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.703732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-serving-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.703818 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.704009 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ddb369-1909-4d63-a0c0-b250490992c0-config\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.704840 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-audit\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.705808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-image-import-ca\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.705992 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.706256 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.707683 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.708584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-encryption-config\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.708838 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.710055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-etcd-client\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.711880 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.712025 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-serving-cert\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.713802 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ddb369-1909-4d63-a0c0-b250490992c0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.718900 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.722342 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0edb65-3bcf-484f-9707-d8124df1ec88-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.722553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwlw\" (UniqueName: \"kubernetes.io/projected/158a5836-f175-4da3-b22d-6a3130a89d30-kube-api-access-ngwlw\") pod \"authentication-operator-69f744f599-xh9n5\" (UID: \"158a5836-f175-4da3-b22d-6a3130a89d30\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.727475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s2kq\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-kube-api-access-6s2kq\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.749147 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c94c\" (UniqueName: \"kubernetes.io/projected/c0da800f-a7ca-4d0e-89bb-96673854969e-kube-api-access-5c94c\") pod \"migrator-59844c95c7-6572z\" (UID: \"c0da800f-a7ca-4d0e-89bb-96673854969e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.764111 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2cdd\" (UniqueName: \"kubernetes.io/projected/99efba52-bc27-49d7-8efb-154b6e3787a9-kube-api-access-f2cdd\") pod \"cluster-samples-operator-665b6dd947-qs72s\" (UID: \"99efba52-bc27-49d7-8efb-154b6e3787a9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.785221 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6a9601-6689-435b-aca1-256a0c3c07fb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltwbb\" (UID: \"6b6a9601-6689-435b-aca1-256a0c3c07fb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.803835 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.804666 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.304649401 +0000 UTC m=+226.796383726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.811836 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.818956 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z955v\" (UniqueName: \"kubernetes.io/projected/9c5f567e-b38f-44a0-b1fd-1a96857e811f-kube-api-access-z955v\") pod \"router-default-5444994796-xl5d7\" (UID: \"9c5f567e-b38f-44a0-b1fd-1a96857e811f\") " pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.819293 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.838050 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.838701 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.842140 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.847841 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.868660 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.872303 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.890547 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.896986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa4b5542-dc36-4c93-88e5-a080729b94ae-config-volume\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.905089 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.905381 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.905530 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.405506137 +0000 UTC m=+226.897240462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.905918 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:26 crc kubenswrapper[4733]: E0318 10:16:26.906258 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.406250841 +0000 UTC m=+226.897985166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.908834 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.916432 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa4b5542-dc36-4c93-88e5-a080729b94ae-metrics-tls\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.924420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.926297 4733 request.go:700] Waited for 1.903494839s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.928065 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.935649 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:26 crc kubenswrapper[4733]: W0318 10:16:26.936526 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2539fca8_3dde_43ed_815c_e837f37dfdd5.slice/crio-7641cfeea867267cd87fc0fa0061dc622336101049c7d0d10a3af6628d0c4a6d WatchSource:0}: Error finding container 7641cfeea867267cd87fc0fa0061dc622336101049c7d0d10a3af6628d0c4a6d: Status 404 returned error can't find the container with id 7641cfeea867267cd87fc0fa0061dc622336101049c7d0d10a3af6628d0c4a6d Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.949537 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.970258 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zztn5"] Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.971516 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.978742 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5979b3d-b4b4-4081-b486-4fcf91f6367c-cert\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.978860 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" Mar 18 10:16:26 crc kubenswrapper[4733]: I0318 10:16:26.988742 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.007281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.007448 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.507422077 +0000 UTC m=+226.999156402 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.007782 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.008067 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.508054408 +0000 UTC m=+226.999788733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.013806 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.023625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-certs\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.028833 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.032689 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.051220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.063287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-node-bootstrap-token\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.092506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqrn\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.111407 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.111965 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.611945101 +0000 UTC m=+227.103679426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.124820 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.131701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lrt\" (UniqueName: \"kubernetes.io/projected/352d0ed5-c43b-431f-bd66-1749ab30d013-kube-api-access-l4lrt\") pod \"apiserver-7bbb656c7d-lsqn4\" (UID: \"352d0ed5-c43b-431f-bd66-1749ab30d013\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.148325 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d48vf\" (UniqueName: \"kubernetes.io/projected/57151941-19ac-4bb5-a93b-b5dfbc88e0d6-kube-api-access-d48vf\") pod \"machine-config-controller-84d6567774-hw7zb\" (UID: \"57151941-19ac-4bb5-a93b-b5dfbc88e0d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.165505 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfbq5\" (UniqueName: \"kubernetes.io/projected/9571ba80-f267-46ed-8d16-e44531cb0ce8-kube-api-access-wfbq5\") pod \"olm-operator-6b444d44fb-2wc5m\" (UID: \"9571ba80-f267-46ed-8d16-e44531cb0ce8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.187756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4clx\" (UniqueName: \"kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx\") pod \"oauth-openshift-558db77b4-n6hmz\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.197909 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.204824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px4sd\" (UniqueName: \"kubernetes.io/projected/d5979b3d-b4b4-4081-b486-4fcf91f6367c-kube-api-access-px4sd\") pod \"ingress-canary-pvlch\" (UID: \"d5979b3d-b4b4-4081-b486-4fcf91f6367c\") " pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.213135 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.215428 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.7154095 +0000 UTC m=+227.207143825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.225444 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.230794 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.234165 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mscv\" (UniqueName: \"kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv\") pod \"marketplace-operator-79b997595-9h9xr\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.238928 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.246893 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2kgf\" (UniqueName: \"kubernetes.io/projected/a39a28f7-1fd2-44f7-8b49-05a0faf1e000-kube-api-access-r2kgf\") pod \"machine-config-server-fnzxw\" (UID: \"a39a28f7-1fd2-44f7-8b49-05a0faf1e000\") " pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.250976 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.252664 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gxcb2"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.268615 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfv2\" (UniqueName: \"kubernetes.io/projected/fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf-kube-api-access-ppfv2\") pod \"openshift-controller-manager-operator-756b6f6bc6-mj46t\" (UID: \"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.287839 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-xh9n5"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.290989 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pvlch" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.295855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fnzxw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.298044 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.301550 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.305158 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.305820 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4fz\" (UniqueName: \"kubernetes.io/projected/ed943a82-ef39-4ebc-9d76-09bb69f3b800-kube-api-access-km4fz\") pod \"csi-hostpathplugin-p4b5s\" (UID: \"ed943a82-ef39-4ebc-9d76-09bb69f3b800\") " pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:27 crc kubenswrapper[4733]: W0318 10:16:27.312759 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod158a5836_f175_4da3_b22d_6a3130a89d30.slice/crio-a8d759096e3def15bd7fc6782fe1ae3d0b85fb4bb74cad2d5fdba68e3116790f WatchSource:0}: Error finding container a8d759096e3def15bd7fc6782fe1ae3d0b85fb4bb74cad2d5fdba68e3116790f: Status 404 returned error can't find the container with id a8d759096e3def15bd7fc6782fe1ae3d0b85fb4bb74cad2d5fdba68e3116790f Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.314293 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.314488 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.814463249 +0000 UTC m=+227.306197574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.314804 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.315247 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.815238174 +0000 UTC m=+227.306972499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: W0318 10:16:27.323013 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0da800f_a7ca_4d0e_89bb_96673854969e.slice/crio-8c23f5367b08d7ec924e81e14494da32936d3eded7ada5c1129339081e4668e4 WatchSource:0}: Error finding container 8c23f5367b08d7ec924e81e14494da32936d3eded7ada5c1129339081e4668e4: Status 404 returned error can't find the container with id 8c23f5367b08d7ec924e81e14494da32936d3eded7ada5c1129339081e4668e4 Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.324531 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngmrr\" (UniqueName: \"kubernetes.io/projected/b405f127-b181-49a1-8205-aafd58d1fa7b-kube-api-access-ngmrr\") pod \"cluster-image-registry-operator-dc59b4c8b-gdmqx\" (UID: \"b405f127-b181-49a1-8205-aafd58d1fa7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.346422 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.350264 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84ddb369-1909-4d63-a0c0-b250490992c0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bzhq6\" (UID: \"84ddb369-1909-4d63-a0c0-b250490992c0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.369570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.370672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2qxh\" (UniqueName: \"kubernetes.io/projected/56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6-kube-api-access-c2qxh\") pod \"apiserver-76f77b778f-xvnwv\" (UID: \"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6\") " pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:27 crc kubenswrapper[4733]: W0318 10:16:27.400135 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda39a28f7_1fd2_44f7_8b49_05a0faf1e000.slice/crio-b8f6e716a5c1af59b334b099ae81e6b4488d512e7cedc4016a21f3ebde2403f2 WatchSource:0}: Error finding container b8f6e716a5c1af59b334b099ae81e6b4488d512e7cedc4016a21f3ebde2403f2: Status 404 returned error can't find the container with id b8f6e716a5c1af59b334b099ae81e6b4488d512e7cedc4016a21f3ebde2403f2 Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.402446 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6nw8\" (UniqueName: \"kubernetes.io/projected/10e64d74-2e25-41fd-a9ad-32a3e74e5c01-kube-api-access-h6nw8\") pod \"dns-operator-744455d44c-vsnq2\" (UID: \"10e64d74-2e25-41fd-a9ad-32a3e74e5c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.404579 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fblf6\" (UniqueName: \"kubernetes.io/projected/4810c2fd-346b-44a0-b985-46d302060373-kube-api-access-fblf6\") pod \"console-operator-58897d9998-lptjf\" (UID: \"4810c2fd-346b-44a0-b985-46d302060373\") " pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.408452 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.415507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.415948 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:27.915932915 +0000 UTC m=+227.407667230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.420838 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.425143 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" event={"ID":"158a5836-f175-4da3-b22d-6a3130a89d30","Type":"ContainerStarted","Data":"a8d759096e3def15bd7fc6782fe1ae3d0b85fb4bb74cad2d5fdba68e3116790f"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.427634 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.428518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xl5d7" event={"ID":"9c5f567e-b38f-44a0-b1fd-1a96857e811f","Type":"ContainerStarted","Data":"6a0821eaf0ace459007c5e1ac4ea5e8e71a9bf775a8c12d7396c4a66aae6f399"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.428556 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-xl5d7" event={"ID":"9c5f567e-b38f-44a0-b1fd-1a96857e811f","Type":"ContainerStarted","Data":"77a44e88ff52ee3e93fedaca9ef3e99488e511df1a94502edd941b321649b22f"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.434585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2whnv\" (UniqueName: \"kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv\") pod \"auto-csr-approver-29563816-4582s\" (UID: \"71a70c3c-d483-43f4-9f54-10978c7f8cc8\") " pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.436005 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gxcb2" event={"ID":"61e27ee7-5eb0-4cc7-a696-85ddd192b171","Type":"ContainerStarted","Data":"c6a2e1b73a77cc11d754771ab069f0ca8832726e42d557307e0ab4ffb83a3bf0"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.445597 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8v244" event={"ID":"f27409fc-b6dd-4573-918b-7b30b3635cc7","Type":"ContainerStarted","Data":"a5e5da5d6249a1112447a42843768f7217f63fd427eb58063240eac26ad5daee"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.448473 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" event={"ID":"99efba52-bc27-49d7-8efb-154b6e3787a9","Type":"ContainerStarted","Data":"d43b8bdc150cb1529e884ad3fa3abe213ef64ab2d2f93043bda6683af7158f68"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.449350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn4vp\" (UniqueName: \"kubernetes.io/projected/3a0400a1-7e6b-4335-8819-586d7a460e3d-kube-api-access-vn4vp\") pod \"packageserver-d55dfcdfc-z8g4f\" (UID: \"3a0400a1-7e6b-4335-8819-586d7a460e3d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.456461 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" event={"ID":"43cea3fb-14f9-4993-a8a9-4618680e8286","Type":"ContainerStarted","Data":"84a76e4c28a1a47f5b3b223ed33580d3157658233ba71b9ee938f5d574b00588"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.456505 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.458834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.470426 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scsd7\" (UniqueName: \"kubernetes.io/projected/53a14d61-5c2c-44b8-b3cb-c8daa23762bf-kube-api-access-scsd7\") pod \"service-ca-9c57cc56f-h5xdn\" (UID: \"53a14d61-5c2c-44b8-b3cb-c8daa23762bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.476102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" event={"ID":"d9572819-3894-4603-bd2b-7c9465bb0067","Type":"ContainerStarted","Data":"60211db626956dc1b7fbfef86dfdf4c016f3357c9749f94b72f116821c89798b"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.479629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" event={"ID":"c0da800f-a7ca-4d0e-89bb-96673854969e","Type":"ContainerStarted","Data":"8c23f5367b08d7ec924e81e14494da32936d3eded7ada5c1129339081e4668e4"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.488390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" event={"ID":"6b6a9601-6689-435b-aca1-256a0c3c07fb","Type":"ContainerStarted","Data":"af40990270173839420767ce10fda62401f1d57ecdf0a8838d88225d26980226"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.496408 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" event={"ID":"2539fca8-3dde-43ed-815c-e837f37dfdd5","Type":"ContainerStarted","Data":"6cd2d6a279ab711c3730b39d682356d39f6069afd8b89fe32581e162caee0dfd"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.496455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" event={"ID":"2539fca8-3dde-43ed-815c-e837f37dfdd5","Type":"ContainerStarted","Data":"7641cfeea867267cd87fc0fa0061dc622336101049c7d0d10a3af6628d0c4a6d"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.497498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fnzxw" event={"ID":"a39a28f7-1fd2-44f7-8b49-05a0faf1e000","Type":"ContainerStarted","Data":"b8f6e716a5c1af59b334b099ae81e6b4488d512e7cedc4016a21f3ebde2403f2"} Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.511339 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7522\" (UniqueName: \"kubernetes.io/projected/ad648fa7-2560-4aa0-8634-05bcbc48916f-kube-api-access-n7522\") pod \"service-ca-operator-777779d784-t95b6\" (UID: \"ad648fa7-2560-4aa0-8634-05bcbc48916f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.516136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grpxc\" (UniqueName: \"kubernetes.io/projected/ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806-kube-api-access-grpxc\") pod \"machine-config-operator-74547568cd-mz68f\" (UID: \"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.517791 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.520477 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.020457248 +0000 UTC m=+227.512191573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.541822 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.541951 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.551586 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.554585 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.558244 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9l2x\" (UniqueName: \"kubernetes.io/projected/87157be2-0fc3-4120-b9b6-d4494ace940a-kube-api-access-j9l2x\") pod \"multus-admission-controller-857f4d67dd-h8kqf\" (UID: \"87157be2-0fc3-4120-b9b6-d4494ace940a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.561047 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.561735 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3be6d75e-e4f8-4d9b-8ed3-9d25632de88c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mxb9q\" (UID: \"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.563675 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjbj\" (UniqueName: \"kubernetes.io/projected/9b0edb65-3bcf-484f-9707-d8124df1ec88-kube-api-access-ncjbj\") pod \"package-server-manager-789f6589d5-kd6gw\" (UID: \"9b0edb65-3bcf-484f-9707-d8124df1ec88\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.577047 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.591707 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zf8\" (UniqueName: \"kubernetes.io/projected/10f3d99e-72fa-4c62-8190-059d7a0effd1-kube-api-access-t4zf8\") pod \"kube-storage-version-migrator-operator-b67b599dd-dxd7p\" (UID: \"10f3d99e-72fa-4c62-8190-059d7a0effd1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.605290 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97ffe185-3f09-44d0-a173-f95bb53c419e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8kv4d\" (UID: \"97ffe185-3f09-44d0-a173-f95bb53c419e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.618958 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.621130 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.121114378 +0000 UTC m=+227.612848703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.624418 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lfsh\" (UniqueName: \"kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh\") pod \"collect-profiles-29563815-tsrs6\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.645547 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kwjh\" (UniqueName: \"kubernetes.io/projected/f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7-kube-api-access-6kwjh\") pod \"control-plane-machine-set-operator-78cbb6b69f-4lbr5\" (UID: \"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.652104 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.660407 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.675330 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.681645 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzlbf\" (UniqueName: \"kubernetes.io/projected/aa4b5542-dc36-4c93-88e5-a080729b94ae-kube-api-access-dzlbf\") pod \"dns-default-hvmrz\" (UID: \"aa4b5542-dc36-4c93-88e5-a080729b94ae\") " pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.689999 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw7q6\" (UniqueName: \"kubernetes.io/projected/4edea753-21f5-44fd-b183-daf03845dcd8-kube-api-access-gw7q6\") pod \"catalog-operator-68c6474976-g686q\" (UID: \"4edea753-21f5-44fd-b183-daf03845dcd8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.696792 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.709346 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll8g5\" (UniqueName: \"kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5\") pod \"controller-manager-879f6c89f-7rr85\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.714839 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.721656 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.722168 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.22215283 +0000 UTC m=+227.713887165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.738960 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.753555 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.766084 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.777673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.825298 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.825759 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.325740574 +0000 UTC m=+227.817474899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.840566 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.845740 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.846014 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pvlch"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.856708 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:27 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:27 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:27 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.856753 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.860640 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-n6hmz"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.891573 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.902660 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.909444 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb"] Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.929970 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:27 crc kubenswrapper[4733]: E0318 10:16:27.930561 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.430537936 +0000 UTC m=+227.922272261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:27 crc kubenswrapper[4733]: I0318 10:16:27.980767 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:27.999262 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.031620 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.032097 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.532081524 +0000 UTC m=+228.023815849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.133764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.134341 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.634317315 +0000 UTC m=+228.126051640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.235134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.235416 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.735394928 +0000 UTC m=+228.227129253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.235752 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.235999 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.735988067 +0000 UTC m=+228.227722382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.288998 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lptjf"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.307302 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vsnq2"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.315347 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.339135 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.339390 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.839354963 +0000 UTC m=+228.331089288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.339634 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.339998 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.839990154 +0000 UTC m=+228.331724479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.441510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.442305 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.942281476 +0000 UTC m=+228.434015801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.446580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.447170 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:28.947158812 +0000 UTC m=+228.438893127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.534257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" event={"ID":"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1","Type":"ContainerStarted","Data":"a4a546ed80545bf50a0d399d05bcd3718be5de86367b6c0e97b326427eeeb776"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.541621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" event={"ID":"6b6a9601-6689-435b-aca1-256a0c3c07fb","Type":"ContainerStarted","Data":"72854d30dcf4277263030dcc04fa8e3b931cb135f23b8f64dd5d658527b9294a"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.541670 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" event={"ID":"6b6a9601-6689-435b-aca1-256a0c3c07fb","Type":"ContainerStarted","Data":"c1a12174bc8f1eb67af8ffffdb1fffbcbcdb362d4ea4dbcf508b0589cbe29f08"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.545361 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" event={"ID":"352d0ed5-c43b-431f-bd66-1749ab30d013","Type":"ContainerStarted","Data":"203bbcbf6e2d03e6169b7db13046fb530cfcba776d38b189faf47c799b7d4c84"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.547847 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.548020 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.047996687 +0000 UTC m=+228.539731012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.548343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.548747 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.048730261 +0000 UTC m=+228.540464586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.554027 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.555869 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gxcb2" event={"ID":"61e27ee7-5eb0-4cc7-a696-85ddd192b171","Type":"ContainerStarted","Data":"0e6c3d3da7414da2e0ca8efb58ca0763564c7f6014da1bcee33997eb225ede97"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.555993 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.562993 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-gxcb2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.563066 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gxcb2" podUID="61e27ee7-5eb0-4cc7-a696-85ddd192b171" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.564351 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563816-4582s"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.566152 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pvlch" event={"ID":"d5979b3d-b4b4-4081-b486-4fcf91f6367c","Type":"ContainerStarted","Data":"113cc6c1a897cb5a44dd78a65cb2ea6a6b8c07d1baccca5b4ff1d285831d52f9"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.569117 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8v244" event={"ID":"f27409fc-b6dd-4573-918b-7b30b3635cc7","Type":"ContainerStarted","Data":"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.588221 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t95b6"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.588281 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lptjf" event={"ID":"4810c2fd-346b-44a0-b985-46d302060373","Type":"ContainerStarted","Data":"424301d2dbd8aeaa7b172e2a161e360ebcd9209429a5cadd81a09d0602e3686c"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.588305 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.594287 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.612628 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" event={"ID":"c0da800f-a7ca-4d0e-89bb-96673854969e","Type":"ContainerStarted","Data":"118ea78d245d5b728468627b8c7ab47e6f832335830c25ed70b25f6dd7a01a25"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.640779 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fnzxw" event={"ID":"a39a28f7-1fd2-44f7-8b49-05a0faf1e000","Type":"ContainerStarted","Data":"30b798fe671c0535b3caa99e9a4d56b7a41c2f648180dfebad30e138508b8aa4"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.651607 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.652731 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.152709927 +0000 UTC m=+228.644444242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.675387 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" event={"ID":"9571ba80-f267-46ed-8d16-e44531cb0ce8","Type":"ContainerStarted","Data":"7509536c23a530965d60a2babd5ba5f501ce7c3e5ae88f80ac6edd7689874c81"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.676560 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" event={"ID":"57151941-19ac-4bb5-a93b-b5dfbc88e0d6","Type":"ContainerStarted","Data":"a711eb4e22c7790a7f75329be3d0673d6babb3f6cd16985bba1d5a761f53ad59"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.690175 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" event={"ID":"99efba52-bc27-49d7-8efb-154b6e3787a9","Type":"ContainerStarted","Data":"6ca6d3f8258ac7f1ba833759822d563e438abce6593a7151642000287f32dcb6"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.693972 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" event={"ID":"d9572819-3894-4603-bd2b-7c9465bb0067","Type":"ContainerStarted","Data":"80055ea6064b163f8ba1a16c665bd9bac35a28e65e0935c42762d28779dcf6cd"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.708013 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" event={"ID":"2539fca8-3dde-43ed-815c-e837f37dfdd5","Type":"ContainerStarted","Data":"ecf92323ff5f1b4df76e0fcb1d32e8f9f8f8594734aaf26d3d3727e3e12ba2b6"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.718879 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" event={"ID":"158a5836-f175-4da3-b22d-6a3130a89d30","Type":"ContainerStarted","Data":"35bffb9ac015b9208a2749f27ac917300045ccb592684109b3caa94bf45c1e9b"} Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.753774 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.755766 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.255750593 +0000 UTC m=+228.747484918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.844921 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:28 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:28 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:28 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.845444 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.859592 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.860004 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.359989058 +0000 UTC m=+228.851723383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.961323 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:28 crc kubenswrapper[4733]: E0318 10:16:28.961756 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.461743942 +0000 UTC m=+228.953478267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.967507 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.985759 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5"] Mar 18 10:16:28 crc kubenswrapper[4733]: I0318 10:16:28.985827 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-xvnwv"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.062712 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.063353 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.563332832 +0000 UTC m=+229.055067157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.063511 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.063789 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.563775796 +0000 UTC m=+229.055510121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.065862 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" podStartSLOduration=171.065838862 podStartE2EDuration="2m51.065838862s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.062256858 +0000 UTC m=+228.553991183" watchObservedRunningTime="2026-03-18 10:16:29.065838862 +0000 UTC m=+228.557573187" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.164574 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.66454133 +0000 UTC m=+229.156275655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.164075 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.168453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.177396 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.6773641 +0000 UTC m=+229.169098425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.181814 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-nbftd" podStartSLOduration=171.181794921 podStartE2EDuration="2m51.181794921s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.180307764 +0000 UTC m=+228.672042089" watchObservedRunningTime="2026-03-18 10:16:29.181794921 +0000 UTC m=+228.673529246" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.269354 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.270121 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.770105156 +0000 UTC m=+229.261839481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.372182 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.372520 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.872500232 +0000 UTC m=+229.364234557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.384223 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7z2vw" podStartSLOduration=171.384205606 podStartE2EDuration="2m51.384205606s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.383523114 +0000 UTC m=+228.875257449" watchObservedRunningTime="2026-03-18 10:16:29.384205606 +0000 UTC m=+228.875939931" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.447484 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4b5s"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.451548 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.461038 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.466811 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" podStartSLOduration=171.466795118 podStartE2EDuration="2m51.466795118s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.466293052 +0000 UTC m=+228.958027377" watchObservedRunningTime="2026-03-18 10:16:29.466795118 +0000 UTC m=+228.958529443" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.476812 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.477248 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:29.977232482 +0000 UTC m=+229.468966807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.491096 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.492629 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.558244 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.571628 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-xl5d7" podStartSLOduration=171.571602391 podStartE2EDuration="2m51.571602391s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.555216756 +0000 UTC m=+229.046951081" watchObservedRunningTime="2026-03-18 10:16:29.571602391 +0000 UTC m=+229.063336716" Mar 18 10:16:29 crc kubenswrapper[4733]: W0318 10:16:29.574821 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ea1a9f_9093_421f_bef3_228352aa65fb.slice/crio-9ae227eb47f15060adbbd96eb5744108b09bf4ae0d948bfba7b04b2c867d1d95 WatchSource:0}: Error finding container 9ae227eb47f15060adbbd96eb5744108b09bf4ae0d948bfba7b04b2c867d1d95: Status 404 returned error can't find the container with id 9ae227eb47f15060adbbd96eb5744108b09bf4ae0d948bfba7b04b2c867d1d95 Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.579085 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.579496 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.079484173 +0000 UTC m=+229.571218498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.590663 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.600707 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-xh9n5" podStartSLOduration=171.600685941 podStartE2EDuration="2m51.600685941s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.599619767 +0000 UTC m=+229.091354092" watchObservedRunningTime="2026-03-18 10:16:29.600685941 +0000 UTC m=+229.092420266" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.626609 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h8kqf"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.642725 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40554: no serving certificate available for the kubelet" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.649720 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h5xdn"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.659684 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.662855 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fnzxw" podStartSLOduration=5.6628391879999995 podStartE2EDuration="5.662839188s" podCreationTimestamp="2026-03-18 10:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.627025724 +0000 UTC m=+229.118760059" watchObservedRunningTime="2026-03-18 10:16:29.662839188 +0000 UTC m=+229.154573513" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.670971 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.672041 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8v244" podStartSLOduration=171.672028932 podStartE2EDuration="2m51.672028932s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.66538943 +0000 UTC m=+229.157123755" watchObservedRunningTime="2026-03-18 10:16:29.672028932 +0000 UTC m=+229.163763257" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.680559 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.681050 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.18103307 +0000 UTC m=+229.672767395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.712227 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.715702 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-gxcb2" podStartSLOduration=171.715663268 podStartE2EDuration="2m51.715663268s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.709233692 +0000 UTC m=+229.200968027" watchObservedRunningTime="2026-03-18 10:16:29.715663268 +0000 UTC m=+229.207397603" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.723483 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hvmrz"] Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.738076 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40556: no serving certificate available for the kubelet" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.746282 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9dd56" podStartSLOduration=171.746254796 podStartE2EDuration="2m51.746254796s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.746247316 +0000 UTC m=+229.237981661" watchObservedRunningTime="2026-03-18 10:16:29.746254796 +0000 UTC m=+229.237989121" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.758213 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" event={"ID":"3a0400a1-7e6b-4335-8819-586d7a460e3d","Type":"ContainerStarted","Data":"c4a9ee364cd2466cb55a3d51888b8c4101c267164bf3b5436f6235972044a23d"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.766080 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" event={"ID":"97ffe185-3f09-44d0-a173-f95bb53c419e","Type":"ContainerStarted","Data":"c46d0d6a694f3a8b4e2de7f5805dda623a4efdfb1d033cf681612d347d2aa0a3"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.788129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.789607 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.289587673 +0000 UTC m=+229.781321998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.791435 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" event={"ID":"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf","Type":"ContainerStarted","Data":"7968cddcf548bf1f43ac4db94f04ec9f35ec3eef36c903d4c32d8745369c8245"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.798723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" event={"ID":"9b0edb65-3bcf-484f-9707-d8124df1ec88","Type":"ContainerStarted","Data":"ffbc1663983c6339c5f1c43fda83f1cfb6b583c431dab23cd21586b747a1455c"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.802524 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" podStartSLOduration=171.802503586 podStartE2EDuration="2m51.802503586s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.794249452 +0000 UTC m=+229.285983777" watchObservedRunningTime="2026-03-18 10:16:29.802503586 +0000 UTC m=+229.294237901" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.828998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" event={"ID":"b405f127-b181-49a1-8205-aafd58d1fa7b","Type":"ContainerStarted","Data":"5fe24ce835535b837457b8d82041843f9c571cab16780054b4369e26fe3d7709"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.829049 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" event={"ID":"b405f127-b181-49a1-8205-aafd58d1fa7b","Type":"ContainerStarted","Data":"235ba7c7a945d007329d95e7f0539da81a2264608d4ac0ddc4e6122ef0b4674b"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.832415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" event={"ID":"ed943a82-ef39-4ebc-9d76-09bb69f3b800","Type":"ContainerStarted","Data":"7c30dd242387191c6740e0ff74bbf53b6106d712df90701a2547298bdbf37b7f"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.833324 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-zztn5" podStartSLOduration=171.833305901 podStartE2EDuration="2m51.833305901s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.830825312 +0000 UTC m=+229.322559637" watchObservedRunningTime="2026-03-18 10:16:29.833305901 +0000 UTC m=+229.325040226" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.835641 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40566: no serving certificate available for the kubelet" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.838681 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" event={"ID":"9571ba80-f267-46ed-8d16-e44531cb0ce8","Type":"ContainerStarted","Data":"184a1347e2d9e7d55a6bb8afbead124af12ccaa51acec4577b15265b8ce508c8"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.842107 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.842131 4733 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2wc5m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.842180 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" podUID="9571ba80-f267-46ed-8d16-e44531cb0ce8" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.844626 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" event={"ID":"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6","Type":"ContainerStarted","Data":"82712470f77bccb95e10909493ae2791b9254e9d32324a113d332c37e6e2d3fa"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.848973 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:29 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:29 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:29 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.849043 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.857364 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563816-4582s" event={"ID":"71a70c3c-d483-43f4-9f54-10978c7f8cc8","Type":"ContainerStarted","Data":"f9019fd1aca4002d61050c62413d5f0b6ff4613e81da7416fe1c8a2924a20e45"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.861392 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" event={"ID":"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c","Type":"ContainerStarted","Data":"1f126ed2ad081a81751bb1d13eee61d744b9ca8e02bc19959b6f61e8304ddd49"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.866726 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" event={"ID":"10e64d74-2e25-41fd-a9ad-32a3e74e5c01","Type":"ContainerStarted","Data":"7ad83af7f3d2affc856d9f1d6fdf699c0d30ad628e2aaa849d3a379d7eac769b"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.866844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" event={"ID":"10e64d74-2e25-41fd-a9ad-32a3e74e5c01","Type":"ContainerStarted","Data":"19c874d28f8aad6daae528555cc1f657cd4fe256c49afabb8796c57d354445a5"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.889649 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pvlch" event={"ID":"d5979b3d-b4b4-4081-b486-4fcf91f6367c","Type":"ContainerStarted","Data":"819d4caa673ff793c5cea8327f5cf458e46f808282d254d0e7ec7b02cebaa2d8"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.890101 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltwbb" podStartSLOduration=171.890075127 podStartE2EDuration="2m51.890075127s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.864881911 +0000 UTC m=+229.356616236" watchObservedRunningTime="2026-03-18 10:16:29.890075127 +0000 UTC m=+229.381809452" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.891061 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.891766 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.391749221 +0000 UTC m=+229.883483536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.899642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" event={"ID":"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7","Type":"ContainerStarted","Data":"5a8e674f13f4d362f5825ede2deb20ef47683fe21530ead2b54068d18e923d72"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.902425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerStarted","Data":"3d92f9fbfa1c8b8490e331060d587d908cf420777497bf90bb4815f3f49e79dd"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.902468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerStarted","Data":"7e583c6a058ccd4e267ac556fbc1ecc397a1e062881c05b38f716c2d4a35947b"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.902891 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.910725 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9h9xr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.910799 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.922731 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" event={"ID":"84ddb369-1909-4d63-a0c0-b250490992c0","Type":"ContainerStarted","Data":"47c628a39b3f5cebbc6098ab5941f0a3a5f4fd2b1f27c116d8289f2beae5970f"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.943624 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40578: no serving certificate available for the kubelet" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.944920 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" podStartSLOduration=171.94488931 podStartE2EDuration="2m51.94488931s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.943962361 +0000 UTC m=+229.435696686" watchObservedRunningTime="2026-03-18 10:16:29.94488931 +0000 UTC m=+229.436623635" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.946300 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gdmqx" podStartSLOduration=171.946292025 podStartE2EDuration="2m51.946292025s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:29.907292568 +0000 UTC m=+229.399026913" watchObservedRunningTime="2026-03-18 10:16:29.946292025 +0000 UTC m=+229.438026350" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.968628 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lptjf" event={"ID":"4810c2fd-346b-44a0-b985-46d302060373","Type":"ContainerStarted","Data":"735d2c405cd183d67cea5e3779b6db05a6c1b4858cd0631b0d3af1260094e44e"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.969923 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.973608 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" event={"ID":"53a14d61-5c2c-44b8-b3cb-c8daa23762bf","Type":"ContainerStarted","Data":"a7322d3ee85a057af8ee5370aa5b68beac85fd2e9f9f76ed7fe65f16fe471252"} Mar 18 10:16:29 crc kubenswrapper[4733]: I0318 10:16:29.995948 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:29 crc kubenswrapper[4733]: E0318 10:16:29.997742 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.497729071 +0000 UTC m=+229.989463396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.002492 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" event={"ID":"34ea1a9f-9093-421f-bef3-228352aa65fb","Type":"ContainerStarted","Data":"9ae227eb47f15060adbbd96eb5744108b09bf4ae0d948bfba7b04b2c867d1d95"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.002541 4733 patch_prober.go:28] interesting pod/console-operator-58897d9998-lptjf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.002583 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lptjf" podUID="4810c2fd-346b-44a0-b985-46d302060373" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.006493 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" event={"ID":"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806","Type":"ContainerStarted","Data":"11401c60ab3a6d8f146de60797f2a2f1f201df5e3acb2ea3c5026e188ee89b7c"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.012145 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" event={"ID":"ad648fa7-2560-4aa0-8634-05bcbc48916f","Type":"ContainerStarted","Data":"4ae7727fa124f9a4a77bef56557f443b2ac6b4ef5dd776d94f91112478e6c6e7"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.022319 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" podStartSLOduration=172.022300197 podStartE2EDuration="2m52.022300197s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.022277816 +0000 UTC m=+229.514012141" watchObservedRunningTime="2026-03-18 10:16:30.022300197 +0000 UTC m=+229.514034522" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.023248 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pvlch" podStartSLOduration=6.023242097 podStartE2EDuration="6.023242097s" podCreationTimestamp="2026-03-18 10:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.002296777 +0000 UTC m=+229.494031102" watchObservedRunningTime="2026-03-18 10:16:30.023242097 +0000 UTC m=+229.514976412" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.037528 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40586: no serving certificate available for the kubelet" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.045714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" event={"ID":"87157be2-0fc3-4120-b9b6-d4494ace940a","Type":"ContainerStarted","Data":"2b43a0d12ed658afd5390df634ceef4336e71a96359cd4aabe1b8017a2b0df25"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.054087 4733 generic.go:334] "Generic (PLEG): container finished" podID="352d0ed5-c43b-431f-bd66-1749ab30d013" containerID="c66b0d6638a9b20b71ffbf84d0f23d56294dbac26257e2ca3a04af6eb34a1ccf" exitCode=0 Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.054150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" event={"ID":"352d0ed5-c43b-431f-bd66-1749ab30d013","Type":"ContainerDied","Data":"c66b0d6638a9b20b71ffbf84d0f23d56294dbac26257e2ca3a04af6eb34a1ccf"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.070767 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" podStartSLOduration=172.070744516 podStartE2EDuration="2m52.070744516s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.065677294 +0000 UTC m=+229.557411619" watchObservedRunningTime="2026-03-18 10:16:30.070744516 +0000 UTC m=+229.562478841" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.071120 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" event={"ID":"57151941-19ac-4bb5-a93b-b5dfbc88e0d6","Type":"ContainerStarted","Data":"38b3b1f22a2a3afbcf93b73f49a0b9a28fa07f5a614dc83f62b7fd5c9b7ffac1"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.078440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" event={"ID":"c0da800f-a7ca-4d0e-89bb-96673854969e","Type":"ContainerStarted","Data":"8864b7b9e0f87345c4e5bb9b84bac1c3dd7621197cee5f9ae21fd5fc4bee7055"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.086488 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qs72s" event={"ID":"99efba52-bc27-49d7-8efb-154b6e3787a9","Type":"ContainerStarted","Data":"f36c5a646fe472dc33637a91fcbb88bf31006bcc9aac430a12510e48916562d6"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.092476 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" event={"ID":"d915f7d2-5b4d-4017-a839-b615a182fafb","Type":"ContainerStarted","Data":"4abe99dbd7bc3b694bde422289cb5e8d4d69c342990c6d29b9ffcb65e8f885f7"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.098583 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.100429 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.600411745 +0000 UTC m=+230.092146070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.116100 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" podStartSLOduration=172.116080677 podStartE2EDuration="2m52.116080677s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.115679634 +0000 UTC m=+229.607413969" watchObservedRunningTime="2026-03-18 10:16:30.116080677 +0000 UTC m=+229.607815002" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.139519 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" event={"ID":"4edea753-21f5-44fd-b183-daf03845dcd8","Type":"ContainerStarted","Data":"b051ef59b46aa4f4ca585c5faa0bd51cc805330540dddec38e277e217f6b2385"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.147871 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lptjf" podStartSLOduration=172.147849913 podStartE2EDuration="2m52.147849913s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.147076668 +0000 UTC m=+229.638810993" watchObservedRunningTime="2026-03-18 10:16:30.147849913 +0000 UTC m=+229.639584238" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.155008 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40590: no serving certificate available for the kubelet" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.167820 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" event={"ID":"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1","Type":"ContainerStarted","Data":"2063ba38b8f338dff7686f6578cd42c9d0c532672eb45f293854b46ba18f0fea"} Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.168684 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-gxcb2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.168709 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gxcb2" podUID="61e27ee7-5eb0-4cc7-a696-85ddd192b171" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.169277 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.201787 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.237045 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.737017945 +0000 UTC m=+230.228752270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.275437 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40592: no serving certificate available for the kubelet" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.296062 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" podStartSLOduration=172.296044933 podStartE2EDuration="2m52.296044933s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.251444077 +0000 UTC m=+229.743178392" watchObservedRunningTime="2026-03-18 10:16:30.296044933 +0000 UTC m=+229.787779258" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.303396 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6572z" podStartSLOduration=172.303387358 podStartE2EDuration="2m52.303387358s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.297654905 +0000 UTC m=+229.789389230" watchObservedRunningTime="2026-03-18 10:16:30.303387358 +0000 UTC m=+229.795121683" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.312393 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.312860 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.812841541 +0000 UTC m=+230.304575866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.313111 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.314495 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.814487683 +0000 UTC m=+230.306222008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.330500 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" podStartSLOduration=172.330479025 podStartE2EDuration="2m52.330479025s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.329568986 +0000 UTC m=+229.821303311" watchObservedRunningTime="2026-03-18 10:16:30.330479025 +0000 UTC m=+229.822213350" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.370237 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" podStartSLOduration=172.370211876 podStartE2EDuration="2m52.370211876s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:30.366801027 +0000 UTC m=+229.858535362" watchObservedRunningTime="2026-03-18 10:16:30.370211876 +0000 UTC m=+229.861946201" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.382979 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40606: no serving certificate available for the kubelet" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.414098 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.414319 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.914276455 +0000 UTC m=+230.406010780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.414494 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.414850 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:30.914832163 +0000 UTC m=+230.406566488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.515489 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.516030 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.01600975 +0000 UTC m=+230.507744075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.617136 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.617932 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.117907649 +0000 UTC m=+230.609641974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.719109 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.719464 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.219433637 +0000 UTC m=+230.711167962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.719730 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.720281 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.220273854 +0000 UTC m=+230.712008179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.821272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.825422 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.325393196 +0000 UTC m=+230.817127521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.851086 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:30 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:30 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:30 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.851162 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:30 crc kubenswrapper[4733]: I0318 10:16:30.926653 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:30 crc kubenswrapper[4733]: E0318 10:16:30.927211 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.427161872 +0000 UTC m=+230.918896197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.028608 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.029203 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.529161754 +0000 UTC m=+231.020896079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.067215 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.092153 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40622: no serving certificate available for the kubelet" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.135130 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.135553 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.635536987 +0000 UTC m=+231.127271312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.219615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" event={"ID":"97ffe185-3f09-44d0-a173-f95bb53c419e","Type":"ContainerStarted","Data":"bb2753716ca08e78d86f779de672cd1596dc670da8dd0801512daa94492f1359"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.226304 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bzhq6" event={"ID":"84ddb369-1909-4d63-a0c0-b250490992c0","Type":"ContainerStarted","Data":"75e60fb3f0f458ee4c62fb018066b15b13329856a88c2a0a0d038e4092ccd720"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.234215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" event={"ID":"10f3d99e-72fa-4c62-8190-059d7a0effd1","Type":"ContainerStarted","Data":"5302f1d354e8e9a2de91a66e848b2bc30e7eb85e32973869e327011ccf471b52"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.234288 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" event={"ID":"10f3d99e-72fa-4c62-8190-059d7a0effd1","Type":"ContainerStarted","Data":"90cf3f6d25ee865494955cc69830e65df1fe45b9b7f73813f62b695dffa405ec"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.235976 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.236653 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.736626661 +0000 UTC m=+231.228360986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.239325 4733 generic.go:334] "Generic (PLEG): container finished" podID="56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6" containerID="6a4e3a2300d0d51d2c8f47aef73b486ee2ec5f3083882676ec91b5dcd4699c4b" exitCode=0 Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.239460 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" event={"ID":"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6","Type":"ContainerDied","Data":"6a4e3a2300d0d51d2c8f47aef73b486ee2ec5f3083882676ec91b5dcd4699c4b"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.272832 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hvmrz" event={"ID":"aa4b5542-dc36-4c93-88e5-a080729b94ae","Type":"ContainerStarted","Data":"6aeac68338b17556e78e0693c265521b8112d378d658532a0beb788af0e2a498"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.273453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hvmrz" event={"ID":"aa4b5542-dc36-4c93-88e5-a080729b94ae","Type":"ContainerStarted","Data":"f28fd3ee37f3b46e37e0643e10dfc0ade0cab785919ea69a9d55380f3f92801a"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.288568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" event={"ID":"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806","Type":"ContainerStarted","Data":"17fd416c7f22590f986e89b3b7da77f9f43747162343171b8e0d88fabbe2b739"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.288639 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" event={"ID":"ec8840c3-e0bd-4cf0-9dd4-87d9ae93b806","Type":"ContainerStarted","Data":"299bbe114466f39c88fc0e927e7d41abfe8761da05c19e5db337ebba724013bc"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.290203 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-848w7" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.294334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" event={"ID":"3a0400a1-7e6b-4335-8819-586d7a460e3d","Type":"ContainerStarted","Data":"9f721d740124d28412f1818fcb188fd9065da5b33f9c191a8e29248b89331523"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.295449 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.300427 4733 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-z8g4f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.300489 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" podUID="3a0400a1-7e6b-4335-8819-586d7a460e3d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.323770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t95b6" event={"ID":"ad648fa7-2560-4aa0-8634-05bcbc48916f","Type":"ContainerStarted","Data":"f271cbb98f96b3dc9e2b032327eccbc87d87a011c3c9b9e24fef1ee6ebfc6652"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.335987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" event={"ID":"34ea1a9f-9093-421f-bef3-228352aa65fb","Type":"ContainerStarted","Data":"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.336900 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.338055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.339991 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.839966316 +0000 UTC m=+231.331700641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.343116 4733 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7rr85 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.343372 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.356411 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4lbr5" event={"ID":"f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7","Type":"ContainerStarted","Data":"b1105779c36b18b11083ade64d7fa22cfa35c4260c2e8d7fff9f833daff1fb14"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.368378 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" event={"ID":"fe90c8a8-c79a-4ed5-bec1-5ea07fbad5cf","Type":"ContainerStarted","Data":"f9d06b9c4e750bd26bf32a4044dfa96fbe08e2c0e576b45953de970b2906834a"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.395882 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" event={"ID":"3be6d75e-e4f8-4d9b-8ed3-9d25632de88c","Type":"ContainerStarted","Data":"83e067e82308fb0fa37c1329e5952542a9cf5a80f17058f1f5f162bce23dc702"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.424680 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hw7zb" event={"ID":"57151941-19ac-4bb5-a93b-b5dfbc88e0d6","Type":"ContainerStarted","Data":"4be9612a0415c389c710939bcff2afe74ad74bc4f0a20b0a6cbc8607ac4f5a0c"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.442896 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.443066 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.943040604 +0000 UTC m=+231.434774929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.443280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.444761 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:31.944753428 +0000 UTC m=+231.436487753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.461147 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" event={"ID":"10e64d74-2e25-41fd-a9ad-32a3e74e5c01","Type":"ContainerStarted","Data":"15a437c1a993be4ad55457ce7627e92e26a819579687176c93eacbf55e8fea4b"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.476375 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" event={"ID":"d915f7d2-5b4d-4017-a839-b615a182fafb","Type":"ContainerStarted","Data":"6068780e861c95e2a5524c6995b5943bf2eb924f4e716f49bfa978772d8dc58d"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.513162 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" event={"ID":"352d0ed5-c43b-431f-bd66-1749ab30d013","Type":"ContainerStarted","Data":"6a709b328a9b33a56f3bd83964c1b7d7d9b38767e78a50508642cd2cb1b3e3e3"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.531480 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" event={"ID":"4edea753-21f5-44fd-b183-daf03845dcd8","Type":"ContainerStarted","Data":"43e71689f4fab8e9f651663bc2478910fb3f8ac94b0e8ba1b81778232bce284c"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.532918 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.542532 4733 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-g686q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.542592 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" podUID="4edea753-21f5-44fd-b183-daf03845dcd8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.544944 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.546265 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.046245315 +0000 UTC m=+231.537979640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.548817 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" event={"ID":"87157be2-0fc3-4120-b9b6-d4494ace940a","Type":"ContainerStarted","Data":"c4b3ec47f6207612c1d201c727747846b79e1d7aa74dde8f58a8c4d4d9bdafd6"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.548857 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" event={"ID":"87157be2-0fc3-4120-b9b6-d4494ace940a","Type":"ContainerStarted","Data":"a5688395d385853aec18579f7e1a245d67a8328ab4238c4893b9955e24c11802"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.568998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" event={"ID":"9b0edb65-3bcf-484f-9707-d8124df1ec88","Type":"ContainerStarted","Data":"d3395f5c56bba0e16edb19611a603f0877a770333d6a8dd11f8b6322718eafec"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.569049 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" event={"ID":"9b0edb65-3bcf-484f-9707-d8124df1ec88","Type":"ContainerStarted","Data":"661d9316b3add46a7098ad8b08dc2f8d0026c9f764dbf292a94aa781a7beff3d"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.573745 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.586403 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" event={"ID":"53a14d61-5c2c-44b8-b3cb-c8daa23762bf","Type":"ContainerStarted","Data":"808f7dc813d586a3560911f32dd518a52d54492e86216f58c1d8de30be696b25"} Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.600645 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9h9xr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.600709 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.601143 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.647419 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.652857 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.152844625 +0000 UTC m=+231.644578950 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.749122 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.750302 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.250276122 +0000 UTC m=+231.742010447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.762388 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" podStartSLOduration=173.762356988 podStartE2EDuration="2m53.762356988s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.723710702 +0000 UTC m=+231.215445027" watchObservedRunningTime="2026-03-18 10:16:31.762356988 +0000 UTC m=+231.254091303" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.764660 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mz68f" podStartSLOduration=173.764651651 podStartE2EDuration="2m53.764651651s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.763581057 +0000 UTC m=+231.255315382" watchObservedRunningTime="2026-03-18 10:16:31.764651651 +0000 UTC m=+231.256385976" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.793822 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-h5xdn" podStartSLOduration=173.793806074 podStartE2EDuration="2m53.793806074s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.791466289 +0000 UTC m=+231.283200614" watchObservedRunningTime="2026-03-18 10:16:31.793806074 +0000 UTC m=+231.285540399" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.840722 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vsnq2" podStartSLOduration=173.840699154 podStartE2EDuration="2m53.840699154s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.815488588 +0000 UTC m=+231.307222913" watchObservedRunningTime="2026-03-18 10:16:31.840699154 +0000 UTC m=+231.332433479" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.851119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.851627 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.351610983 +0000 UTC m=+231.843345308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.861599 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:31 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:31 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:31 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.861730 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.888715 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-h8kqf" podStartSLOduration=173.888689059 podStartE2EDuration="2m53.888689059s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.883636077 +0000 UTC m=+231.375370402" watchObservedRunningTime="2026-03-18 10:16:31.888689059 +0000 UTC m=+231.380423384" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.894331 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" podStartSLOduration=173.894304859 podStartE2EDuration="2m53.894304859s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.844398542 +0000 UTC m=+231.336132867" watchObservedRunningTime="2026-03-18 10:16:31.894304859 +0000 UTC m=+231.386039184" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.953814 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:31 crc kubenswrapper[4733]: E0318 10:16:31.954305 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.454289647 +0000 UTC m=+231.946023972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.955797 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8kv4d" podStartSLOduration=173.955780725 podStartE2EDuration="2m53.955780725s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.908588486 +0000 UTC m=+231.400322821" watchObservedRunningTime="2026-03-18 10:16:31.955780725 +0000 UTC m=+231.447515040" Mar 18 10:16:31 crc kubenswrapper[4733]: I0318 10:16:31.991170 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mxb9q" podStartSLOduration=173.991142256 podStartE2EDuration="2m53.991142256s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:31.989612307 +0000 UTC m=+231.481346642" watchObservedRunningTime="2026-03-18 10:16:31.991142256 +0000 UTC m=+231.482876581" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.021381 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mj46t" podStartSLOduration=174.021355973 podStartE2EDuration="2m54.021355973s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.018738719 +0000 UTC m=+231.510473044" watchObservedRunningTime="2026-03-18 10:16:32.021355973 +0000 UTC m=+231.513090298" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.056466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.057436 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.557420146 +0000 UTC m=+232.049154461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.114722 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" podStartSLOduration=174.114707299 podStartE2EDuration="2m54.114707299s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.108550532 +0000 UTC m=+231.600284857" watchObservedRunningTime="2026-03-18 10:16:32.114707299 +0000 UTC m=+231.606441624" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.161329 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.162000 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.661982731 +0000 UTC m=+232.153717046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.236497 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.236986 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.266685 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.267262 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.767238828 +0000 UTC m=+232.258973153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.270797 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" podStartSLOduration=92.270771631 podStartE2EDuration="1m32.270771631s" podCreationTimestamp="2026-03-18 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.26793437 +0000 UTC m=+231.759668715" watchObservedRunningTime="2026-03-18 10:16:32.270771631 +0000 UTC m=+231.762505956" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.357247 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dxd7p" podStartSLOduration=174.357221587 podStartE2EDuration="2m54.357221587s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.317072422 +0000 UTC m=+231.808806747" watchObservedRunningTime="2026-03-18 10:16:32.357221587 +0000 UTC m=+231.848955902" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.370945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.371305 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.871287196 +0000 UTC m=+232.363021521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.389660 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" podStartSLOduration=174.389633853 podStartE2EDuration="2m54.389633853s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.389067495 +0000 UTC m=+231.880801820" watchObservedRunningTime="2026-03-18 10:16:32.389633853 +0000 UTC m=+231.881368178" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.430539 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" podStartSLOduration=174.430524061 podStartE2EDuration="2m54.430524061s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.428877829 +0000 UTC m=+231.920612154" watchObservedRunningTime="2026-03-18 10:16:32.430524061 +0000 UTC m=+231.922258386" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.473916 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.474308 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:32.974295731 +0000 UTC m=+232.466030056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.499505 4733 ???:1] "http: TLS handshake error from 192.168.126.11:40628: no serving certificate available for the kubelet" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.576031 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.577006 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.076982556 +0000 UTC m=+232.568716881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.589770 4733 patch_prober.go:28] interesting pod/console-operator-58897d9998-lptjf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.590041 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lptjf" podUID="4810c2fd-346b-44a0-b985-46d302060373" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.631921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" event={"ID":"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6","Type":"ContainerStarted","Data":"749934ceee0a98d547c6ab88a5257bf4a284fe859a9a2040c67a6910c663a176"} Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.632431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" event={"ID":"56f4298d-e4ce-42a3-a0dc-d9c94b84dfe6","Type":"ContainerStarted","Data":"293173a5534f294877b5577b8f8b99df7629cfb86013c76769749a28770ec93d"} Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.654895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hvmrz" event={"ID":"aa4b5542-dc36-4c93-88e5-a080729b94ae","Type":"ContainerStarted","Data":"442de960ed251c6b8645a6adcae9131d7e11d33e7e1b5f646f14c0f5f4d67fde"} Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.655966 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.681100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.684361 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.184344371 +0000 UTC m=+232.676078696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.685420 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" event={"ID":"ed943a82-ef39-4ebc-9d76-09bb69f3b800","Type":"ContainerStarted","Data":"5569db947d2a68531f9a1c2042230297737a979a298e0604d72073782235d034"} Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.686554 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9h9xr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.686632 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.693047 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.694584 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hvmrz" podStartSLOduration=8.694574308 podStartE2EDuration="8.694574308s" podCreationTimestamp="2026-03-18 10:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.691760118 +0000 UTC m=+232.183494443" watchObservedRunningTime="2026-03-18 10:16:32.694574308 +0000 UTC m=+232.186308633" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.696933 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" podStartSLOduration=174.696923063 podStartE2EDuration="2m54.696923063s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:32.670605711 +0000 UTC m=+232.162340036" watchObservedRunningTime="2026-03-18 10:16:32.696923063 +0000 UTC m=+232.188657388" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.709091 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lptjf" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.709405 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g686q" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.764127 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z8g4f" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.783875 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.786563 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.28653698 +0000 UTC m=+232.778271305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.850825 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:32 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:32 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:32 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.850908 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.886372 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.887102 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.387083696 +0000 UTC m=+232.878818021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.989991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.990513 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.490470693 +0000 UTC m=+232.982205018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:32 crc kubenswrapper[4733]: I0318 10:16:32.990934 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:32 crc kubenswrapper[4733]: E0318 10:16:32.991577 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.491567228 +0000 UTC m=+232.983301553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.054763 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.092251 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.092769 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.592742685 +0000 UTC m=+233.084477010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.092947 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.093599 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.593580211 +0000 UTC m=+233.085314726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.194025 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.194224 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.69419688 +0000 UTC m=+233.185931205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.194671 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.194986 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.694974825 +0000 UTC m=+233.186709150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.296308 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.296652 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.796586894 +0000 UTC m=+233.288321219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.297169 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.297694 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.797685959 +0000 UTC m=+233.289420284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.398136 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.398331 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.898303178 +0000 UTC m=+233.390037503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.398464 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.398866 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:33.898850785 +0000 UTC m=+233.390585110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.499962 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.500884 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.000841198 +0000 UTC m=+233.492575523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.500963 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.501348 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.001332883 +0000 UTC m=+233.493067208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.544563 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.565096 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.565764 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" podUID="25331c44-b639-46f7-8a7f-6f62f8779e2b" containerName="route-controller-manager" containerID="cri-o://3c09df7a275938153d455f147ffe12eff185edea72f1d8646898b9ba5ba684d5" gracePeriod=30 Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.602665 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.603140 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.103118029 +0000 UTC m=+233.594852354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.704243 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" event={"ID":"ed943a82-ef39-4ebc-9d76-09bb69f3b800","Type":"ContainerStarted","Data":"8cb1ca64aafbb94fb628a8e324049ab71e0109fbe0ae58bb9968f02da229b684"} Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.704899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.705365 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.20535277 +0000 UTC m=+233.697087095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.718220 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsqn4" Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.806276 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.807456 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.307410934 +0000 UTC m=+233.799145259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.855843 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:33 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:33 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:33 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.855916 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:33 crc kubenswrapper[4733]: I0318 10:16:33.908475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:33 crc kubenswrapper[4733]: E0318 10:16:33.908900 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.40888337 +0000 UTC m=+233.900617695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.009727 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.010237 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.510207972 +0000 UTC m=+234.001942297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.096336 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.097254 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.109143 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.111983 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.112455 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.612422151 +0000 UTC m=+234.104156476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.164762 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.164816 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.165663 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.172037 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.194172 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.213973 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214355 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j4k9\" (UniqueName: \"kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.214435 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.714401423 +0000 UTC m=+234.206135748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214533 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214583 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214647 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.214832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7hv7\" (UniqueName: \"kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320374 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320510 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320528 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7hv7\" (UniqueName: \"kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320548 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.320574 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j4k9\" (UniqueName: \"kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.321690 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.322053 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.822042607 +0000 UTC m=+234.313776932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.322129 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.322419 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.322506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.328388 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.331085 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.342885 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.359800 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7hv7\" (UniqueName: \"kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7\") pod \"certified-operators-rls2r\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.362506 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j4k9\" (UniqueName: \"kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9\") pod \"community-operators-f92nl\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.421759 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.421918 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.921891951 +0000 UTC m=+234.413626276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.421958 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.422010 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5vpf\" (UniqueName: \"kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.422041 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.422081 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.422392 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:34.922381256 +0000 UTC m=+234.414115581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.427658 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.501498 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.525548 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.525902 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.526121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5vpf\" (UniqueName: \"kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.526159 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.526226 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.526596 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.526686 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.526761 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.026746845 +0000 UTC m=+234.518481170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.527414 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.568340 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.580898 4733 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.582537 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5vpf\" (UniqueName: \"kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf\") pod \"certified-operators-w7rrs\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.636316 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.636402 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs5rx\" (UniqueName: \"kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.636433 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.636820 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.136804125 +0000 UTC m=+234.628538450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.637035 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.722273 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.727636 4733 generic.go:334] "Generic (PLEG): container finished" podID="25331c44-b639-46f7-8a7f-6f62f8779e2b" containerID="3c09df7a275938153d455f147ffe12eff185edea72f1d8646898b9ba5ba684d5" exitCode=0 Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.727739 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" event={"ID":"25331c44-b639-46f7-8a7f-6f62f8779e2b","Type":"ContainerDied","Data":"3c09df7a275938153d455f147ffe12eff185edea72f1d8646898b9ba5ba684d5"} Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.735568 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerName="controller-manager" containerID="cri-o://03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23" gracePeriod=30 Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.735732 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" event={"ID":"ed943a82-ef39-4ebc-9d76-09bb69f3b800","Type":"ContainerStarted","Data":"9e583c49869d66f8518955622944a13bd3b0b124a8088e3d47f4120cdd1db2a7"} Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.735766 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" event={"ID":"ed943a82-ef39-4ebc-9d76-09bb69f3b800","Type":"ContainerStarted","Data":"54d261c692de36f53168ef4d6dffbad6c0b15a9ca777692c089156d4c4a063d1"} Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.738819 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.739179 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.739278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.739326 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs5rx\" (UniqueName: \"kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.739833 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.23981108 +0000 UTC m=+234.731545405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.740313 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.740550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.768707 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs5rx\" (UniqueName: \"kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx\") pod \"community-operators-gmw2d\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.855117 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.856549 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.858586 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.358567629 +0000 UTC m=+234.850301954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.867801 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:34 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:34 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:34 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.867856 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.956609 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:34 crc kubenswrapper[4733]: E0318 10:16:34.957087 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.45706341 +0000 UTC m=+234.948797735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.957313 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:34 crc kubenswrapper[4733]: I0318 10:16:34.990856 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-p4b5s" podStartSLOduration=10.9908261 podStartE2EDuration="10.9908261s" podCreationTimestamp="2026-03-18 10:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:34.789867221 +0000 UTC m=+234.281601546" watchObservedRunningTime="2026-03-18 10:16:34.9908261 +0000 UTC m=+234.482560425" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.058150 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert\") pod \"25331c44-b639-46f7-8a7f-6f62f8779e2b\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.058260 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config\") pod \"25331c44-b639-46f7-8a7f-6f62f8779e2b\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.058379 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzj8w\" (UniqueName: \"kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w\") pod \"25331c44-b639-46f7-8a7f-6f62f8779e2b\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.059450 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca\") pod \"25331c44-b639-46f7-8a7f-6f62f8779e2b\" (UID: \"25331c44-b639-46f7-8a7f-6f62f8779e2b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.059566 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config" (OuterVolumeSpecName: "config") pod "25331c44-b639-46f7-8a7f-6f62f8779e2b" (UID: "25331c44-b639-46f7-8a7f-6f62f8779e2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.060232 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca" (OuterVolumeSpecName: "client-ca") pod "25331c44-b639-46f7-8a7f-6f62f8779e2b" (UID: "25331c44-b639-46f7-8a7f-6f62f8779e2b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.060415 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.060614 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.060639 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25331c44-b639-46f7-8a7f-6f62f8779e2b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.064173 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.564149355 +0000 UTC m=+235.055883680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.069111 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "25331c44-b639-46f7-8a7f-6f62f8779e2b" (UID: "25331c44-b639-46f7-8a7f-6f62f8779e2b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.071562 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w" (OuterVolumeSpecName: "kube-api-access-rzj8w") pod "25331c44-b639-46f7-8a7f-6f62f8779e2b" (UID: "25331c44-b639-46f7-8a7f-6f62f8779e2b"). InnerVolumeSpecName "kube-api-access-rzj8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.097746 4733 ???:1] "http: TLS handshake error from 192.168.126.11:33538: no serving certificate available for the kubelet" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.163254 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.163646 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25331c44-b639-46f7-8a7f-6f62f8779e2b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.163661 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzj8w\" (UniqueName: \"kubernetes.io/projected/25331c44-b639-46f7-8a7f-6f62f8779e2b-kube-api-access-rzj8w\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.163747 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.66372416 +0000 UTC m=+235.155458485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.168126 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.194967 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.195344 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25331c44-b639-46f7-8a7f-6f62f8779e2b" containerName="route-controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.195423 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="25331c44-b639-46f7-8a7f-6f62f8779e2b" containerName="route-controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.195598 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="25331c44-b639-46f7-8a7f-6f62f8779e2b" containerName="route-controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.196048 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.196125 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.196271 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.200215 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.200338 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.265008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.265054 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.265086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.265501 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.765487836 +0000 UTC m=+235.257222161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.296763 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.333143 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.360746 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.366863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config\") pod \"34ea1a9f-9093-421f-bef3-228352aa65fb\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.367214 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll8g5\" (UniqueName: \"kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5\") pod \"34ea1a9f-9093-421f-bef3-228352aa65fb\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.367418 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert\") pod \"34ea1a9f-9093-421f-bef3-228352aa65fb\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.367533 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles\") pod \"34ea1a9f-9093-421f-bef3-228352aa65fb\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.369765 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.370980 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca\") pod \"34ea1a9f-9093-421f-bef3-228352aa65fb\" (UID: \"34ea1a9f-9093-421f-bef3-228352aa65fb\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.371460 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config" (OuterVolumeSpecName: "config") pod "34ea1a9f-9093-421f-bef3-228352aa65fb" (UID: "34ea1a9f-9093-421f-bef3-228352aa65fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.373706 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "34ea1a9f-9093-421f-bef3-228352aa65fb" (UID: "34ea1a9f-9093-421f-bef3-228352aa65fb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.375614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca" (OuterVolumeSpecName: "client-ca") pod "34ea1a9f-9093-421f-bef3-228352aa65fb" (UID: "34ea1a9f-9093-421f-bef3-228352aa65fb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.376098 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5" (OuterVolumeSpecName: "kube-api-access-ll8g5") pod "34ea1a9f-9093-421f-bef3-228352aa65fb" (UID: "34ea1a9f-9093-421f-bef3-228352aa65fb"). InnerVolumeSpecName "kube-api-access-ll8g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.386902 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.886852968 +0000 UTC m=+235.378587293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387032 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387080 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387136 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387370 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387387 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll8g5\" (UniqueName: \"kubernetes.io/projected/34ea1a9f-9093-421f-bef3-228352aa65fb-kube-api-access-ll8g5\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387399 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387412 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34ea1a9f-9093-421f-bef3-228352aa65fb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.387778 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 10:16:35.887770307 +0000 UTC m=+235.379504632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nwhtg" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.387813 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.392309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "34ea1a9f-9093-421f-bef3-228352aa65fb" (UID: "34ea1a9f-9093-421f-bef3-228352aa65fb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.403281 4733 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-18T10:16:34.580923168Z","Handler":null,"Name":""} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.421731 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.422168 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerName="controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.422201 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerName="controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.422409 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerName="controller-manager" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.423032 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.428593 4733 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.428650 4733 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.435745 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.437419 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.488527 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.488972 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34ea1a9f-9093-421f-bef3-228352aa65fb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.492865 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.531816 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.590927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.591134 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.591228 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.591298 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmg68\" (UniqueName: \"kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.591488 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.599489 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.599544 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.630505 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nwhtg\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.693354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.693440 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.693496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.693533 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmg68\" (UniqueName: \"kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.695127 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.696824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.700347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.709733 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmg68\" (UniqueName: \"kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68\") pod \"route-controller-manager-6946475f8-lnppg\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.722782 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.741571 4733 generic.go:334] "Generic (PLEG): container finished" podID="34ea1a9f-9093-421f-bef3-228352aa65fb" containerID="03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23" exitCode=0 Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.741683 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.741656 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" event={"ID":"34ea1a9f-9093-421f-bef3-228352aa65fb","Type":"ContainerDied","Data":"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.742045 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7rr85" event={"ID":"34ea1a9f-9093-421f-bef3-228352aa65fb","Type":"ContainerDied","Data":"9ae227eb47f15060adbbd96eb5744108b09bf4ae0d948bfba7b04b2c867d1d95"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.742084 4733 scope.go:117] "RemoveContainer" containerID="03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.744322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerStarted","Data":"a2b42f75b17ecdce018f92ac6406accaeca335b14c1245cfd417767d5e5802c4"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.747600 4733 generic.go:334] "Generic (PLEG): container finished" podID="92996997-080b-42c9-bc2c-19c2e68db896" containerID="a9bf744158dbc316b120322e1385bd5232386e738d2db0f1d91d2ac7d8a7ad1a" exitCode=0 Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.747859 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerDied","Data":"a9bf744158dbc316b120322e1385bd5232386e738d2db0f1d91d2ac7d8a7ad1a"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.747890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerStarted","Data":"448a9e96bdf06f234c1da361f4be5cda2d36bf670a134ff4f206711028d80cac"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.750528 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerStarted","Data":"bf716e26a7a1e4408c9cf17e7366833bdc30d38efd823adf2eb5d92d8a80e381"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.755441 4733 generic.go:334] "Generic (PLEG): container finished" podID="527056ad-4daf-4dd5-9e31-887d55be0336" containerID="83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274" exitCode=0 Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.755587 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerDied","Data":"83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.756000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerStarted","Data":"5d9e5dab0932c3cd3cd8b8f12fa8d0d49db59eddcefaa706bd16f11d86be1eac"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.779789 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.780384 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95" event={"ID":"25331c44-b639-46f7-8a7f-6f62f8779e2b","Type":"ContainerDied","Data":"0a4e2b2140bacea055efd9eb333f7f7f1da7235e623090af40eaf58bc070ecb2"} Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.813210 4733 scope.go:117] "RemoveContainer" containerID="03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23" Mar 18 10:16:35 crc kubenswrapper[4733]: E0318 10:16:35.817109 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23\": container with ID starting with 03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23 not found: ID does not exist" containerID="03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.817176 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23"} err="failed to get container status \"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23\": rpc error: code = NotFound desc = could not find container \"03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23\": container with ID starting with 03eb463e4dd483e49a80db7e81b412b8d031879512dded00ef00cc215cf14f23 not found: ID does not exist" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.817243 4733 scope.go:117] "RemoveContainer" containerID="3c09df7a275938153d455f147ffe12eff185edea72f1d8646898b9ba5ba684d5" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.830508 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.849980 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:35 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:35 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:35 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.850039 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.861410 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7rr85"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.877353 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.880264 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m5k95"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.891121 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.904742 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.908080 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.910464 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.917058 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.970507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.998269 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.998347 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:35 crc kubenswrapper[4733]: I0318 10:16:35.998418 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5frtf\" (UniqueName: \"kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.101099 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.102923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.103015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5frtf\" (UniqueName: \"kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.102476 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.103709 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.130021 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5frtf\" (UniqueName: \"kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf\") pod \"redhat-marketplace-jb86w\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.222851 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.242551 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.305869 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.306940 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.319211 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.408051 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.408678 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r2vl\" (UniqueName: \"kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.408711 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.490329 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.509656 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.509729 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r2vl\" (UniqueName: \"kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.509764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.510278 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.510329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: W0318 10:16:36.526784 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1954584_f4d6_467a_8b0c_9db32f9e385c.slice/crio-f9b9ef0617ed480f3b6667ef58bad6aa0295e3670f7f52f36e961678f655af59 WatchSource:0}: Error finding container f9b9ef0617ed480f3b6667ef58bad6aa0295e3670f7f52f36e961678f655af59: Status 404 returned error can't find the container with id f9b9ef0617ed480f3b6667ef58bad6aa0295e3670f7f52f36e961678f655af59 Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.538674 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r2vl\" (UniqueName: \"kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl\") pod \"redhat-marketplace-f82xf\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.574121 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.632228 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.798402 4733 generic.go:334] "Generic (PLEG): container finished" podID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerID="17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266" exitCode=0 Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.798999 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerDied","Data":"17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.808292 4733 generic.go:334] "Generic (PLEG): container finished" podID="02cd6358-355c-4db8-b0f7-2528618602ff" containerID="0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed" exitCode=0 Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.808745 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerDied","Data":"0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.819353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerStarted","Data":"152bb2d9d2d5d61c127ef6162804e32f4f4e993fb3a1aa90d7238cb79aedf035"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.819940 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-gxcb2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.819991 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gxcb2" podUID="61e27ee7-5eb0-4cc7-a696-85ddd192b171" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.820644 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-gxcb2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.820670 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-gxcb2" podUID="61e27ee7-5eb0-4cc7-a696-85ddd192b171" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.827888 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" event={"ID":"d1954584-f4d6-467a-8b0c-9db32f9e385c","Type":"ContainerStarted","Data":"bb3b81003e39ecbea213f0c0b02b7e8dae8e9507c2c01e812e0dce4d4f5c71d4"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.827960 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" event={"ID":"d1954584-f4d6-467a-8b0c-9db32f9e385c","Type":"ContainerStarted","Data":"f9b9ef0617ed480f3b6667ef58bad6aa0295e3670f7f52f36e961678f655af59"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.829873 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.840739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.871553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75f42e15-d1dc-4edf-8f2e-daef04ccc601","Type":"ContainerStarted","Data":"e35828e9f81ba55c9c2c8d38d1a7b2cf11a4f98596d388b277800589516f0e19"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.871603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75f42e15-d1dc-4edf-8f2e-daef04ccc601","Type":"ContainerStarted","Data":"d4949220d6469d09f06a115f37de0df57649271ca132ce8e25268effeb41c8de"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.874366 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:36 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:36 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:36 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.874426 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.895348 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" event={"ID":"7b5dc098-4a15-429b-8243-1ac75ce2e0c1","Type":"ContainerStarted","Data":"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.895413 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" event={"ID":"7b5dc098-4a15-429b-8243-1ac75ce2e0c1","Type":"ContainerStarted","Data":"bf9beab436bdff3f99c6c06c629fb5de1f2bcd079250aacd7d55627140dc6e11"} Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.896217 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.906543 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.907110 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.910876 4733 patch_prober.go:28] interesting pod/console-f9d7485db-8v244 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.910933 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8v244" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.919207 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" podStartSLOduration=2.919164463 podStartE2EDuration="2.919164463s" podCreationTimestamp="2026-03-18 10:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:36.894670619 +0000 UTC m=+236.386404944" watchObservedRunningTime="2026-03-18 10:16:36.919164463 +0000 UTC m=+236.410898788" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.925526 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.925512346 podStartE2EDuration="1.925512346s" podCreationTimestamp="2026-03-18 10:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:36.921483597 +0000 UTC m=+236.413217922" watchObservedRunningTime="2026-03-18 10:16:36.925512346 +0000 UTC m=+236.417246671" Mar 18 10:16:36 crc kubenswrapper[4733]: I0318 10:16:36.947762 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" podStartSLOduration=178.947733267 podStartE2EDuration="2m58.947733267s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:36.947484019 +0000 UTC m=+236.439218354" watchObservedRunningTime="2026-03-18 10:16:36.947733267 +0000 UTC m=+236.439467592" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.012879 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:16:37 crc kubenswrapper[4733]: W0318 10:16:37.025439 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc91f12fa_96f0_442a_a3f7_70d56a697839.slice/crio-a8fa061a3aa824aa80f6c1569abe326d18dccd731789c62f81d22de7e9a828d3 WatchSource:0}: Error finding container a8fa061a3aa824aa80f6c1569abe326d18dccd731789c62f81d22de7e9a828d3: Status 404 returned error can't find the container with id a8fa061a3aa824aa80f6c1569abe326d18dccd731789c62f81d22de7e9a828d3 Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.193245 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25331c44-b639-46f7-8a7f-6f62f8779e2b" path="/var/lib/kubelet/pods/25331c44-b639-46f7-8a7f-6f62f8779e2b/volumes" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.193917 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ea1a9f-9093-421f-bef3-228352aa65fb" path="/var/lib/kubelet/pods/34ea1a9f-9093-421f-bef3-228352aa65fb/volumes" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.199406 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.208998 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.304547 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.305736 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.308013 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.324323 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.325445 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.330150 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.330544 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.330760 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.334563 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.379470 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.379542 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.379564 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsw8d\" (UniqueName: \"kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.385579 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.412884 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.413662 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.421147 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.421559 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.421713 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.422143 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.423699 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.423872 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.435365 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.444535 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.483377 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.484555 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.484592 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsw8d\" (UniqueName: \"kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.484642 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.484773 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.485496 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.485641 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.525106 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsw8d\" (UniqueName: \"kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d\") pod \"redhat-operators-hrwxg\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.568484 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.569038 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.588851 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.588969 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589069 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589742 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589796 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvgcj\" (UniqueName: \"kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589837 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.589873 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.590627 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.625770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.637576 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.691434 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.691532 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.691784 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvgcj\" (UniqueName: \"kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.691816 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.691843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.694306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.697210 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.702477 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.704704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.710242 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.713503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.717925 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.720439 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvgcj\" (UniqueName: \"kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj\") pod \"controller-manager-75c9fbd49-qmfds\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.728610 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.740355 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.846551 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:37 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:37 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:37 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.846616 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.896375 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txwcn\" (UniqueName: \"kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.896455 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.896479 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.921290 4733 generic.go:334] "Generic (PLEG): container finished" podID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerID="deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e" exitCode=0 Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.921457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerDied","Data":"deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e"} Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.937829 4733 generic.go:334] "Generic (PLEG): container finished" podID="d915f7d2-5b4d-4017-a839-b615a182fafb" containerID="6068780e861c95e2a5524c6995b5943bf2eb924f4e716f49bfa978772d8dc58d" exitCode=0 Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.937965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" event={"ID":"d915f7d2-5b4d-4017-a839-b615a182fafb","Type":"ContainerDied","Data":"6068780e861c95e2a5524c6995b5943bf2eb924f4e716f49bfa978772d8dc58d"} Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.949421 4733 generic.go:334] "Generic (PLEG): container finished" podID="75f42e15-d1dc-4edf-8f2e-daef04ccc601" containerID="e35828e9f81ba55c9c2c8d38d1a7b2cf11a4f98596d388b277800589516f0e19" exitCode=0 Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.949529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75f42e15-d1dc-4edf-8f2e-daef04ccc601","Type":"ContainerDied","Data":"e35828e9f81ba55c9c2c8d38d1a7b2cf11a4f98596d388b277800589516f0e19"} Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.962917 4733 generic.go:334] "Generic (PLEG): container finished" podID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerID="6da3522bbcdb557467c36bac266a9dafb390a5a917de44dd30de9c3ac03051e1" exitCode=0 Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.963594 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerDied","Data":"6da3522bbcdb557467c36bac266a9dafb390a5a917de44dd30de9c3ac03051e1"} Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.963646 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerStarted","Data":"a8fa061a3aa824aa80f6c1569abe326d18dccd731789c62f81d22de7e9a828d3"} Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.973346 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-xvnwv" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.998353 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txwcn\" (UniqueName: \"kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.998432 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.998456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:37 crc kubenswrapper[4733]: I0318 10:16:37.999388 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.000065 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.049365 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txwcn\" (UniqueName: \"kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn\") pod \"redhat-operators-ll7dp\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.071061 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.252618 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.309133 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.337620 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:16:38 crc kubenswrapper[4733]: W0318 10:16:38.376158 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6503bd00_b300_438d_b10e_27380eaf7d9a.slice/crio-cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e WatchSource:0}: Error finding container cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e: Status 404 returned error can't find the container with id cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e Mar 18 10:16:38 crc kubenswrapper[4733]: W0318 10:16:38.400085 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24e63f18_ff07_40b0_8289_351352d47d0a.slice/crio-4e6f1dbd850522a9734d4ded476f261e3ae7fcfb8b06688f8b915837852ee325 WatchSource:0}: Error finding container 4e6f1dbd850522a9734d4ded476f261e3ae7fcfb8b06688f8b915837852ee325: Status 404 returned error can't find the container with id 4e6f1dbd850522a9734d4ded476f261e3ae7fcfb8b06688f8b915837852ee325 Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.886883 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:38 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:38 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:38 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:38 crc kubenswrapper[4733]: I0318 10:16:38.887566 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.006638 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerStarted","Data":"62aa2aa87c6f58e0a138486db1e0ff0949ce50a5eef4891759673935a2791e3b"} Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.010653 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" event={"ID":"24e63f18-ff07-40b0-8289-351352d47d0a","Type":"ContainerStarted","Data":"4e6f1dbd850522a9734d4ded476f261e3ae7fcfb8b06688f8b915837852ee325"} Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.022023 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6503bd00-b300-438d-b10e-27380eaf7d9a","Type":"ContainerStarted","Data":"cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e"} Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.127961 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.496999 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.513337 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.594924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume\") pod \"d915f7d2-5b4d-4017-a839-b615a182fafb\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.594987 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir\") pod \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.595018 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume\") pod \"d915f7d2-5b4d-4017-a839-b615a182fafb\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.596310 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "75f42e15-d1dc-4edf-8f2e-daef04ccc601" (UID: "75f42e15-d1dc-4edf-8f2e-daef04ccc601"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.596653 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume" (OuterVolumeSpecName: "config-volume") pod "d915f7d2-5b4d-4017-a839-b615a182fafb" (UID: "d915f7d2-5b4d-4017-a839-b615a182fafb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.603018 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d915f7d2-5b4d-4017-a839-b615a182fafb" (UID: "d915f7d2-5b4d-4017-a839-b615a182fafb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.707849 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lfsh\" (UniqueName: \"kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh\") pod \"d915f7d2-5b4d-4017-a839-b615a182fafb\" (UID: \"d915f7d2-5b4d-4017-a839-b615a182fafb\") " Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.707955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access\") pod \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\" (UID: \"75f42e15-d1dc-4edf-8f2e-daef04ccc601\") " Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.708427 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.708447 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d915f7d2-5b4d-4017-a839-b615a182fafb-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.708459 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d915f7d2-5b4d-4017-a839-b615a182fafb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.713323 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh" (OuterVolumeSpecName: "kube-api-access-8lfsh") pod "d915f7d2-5b4d-4017-a839-b615a182fafb" (UID: "d915f7d2-5b4d-4017-a839-b615a182fafb"). InnerVolumeSpecName "kube-api-access-8lfsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.716279 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "75f42e15-d1dc-4edf-8f2e-daef04ccc601" (UID: "75f42e15-d1dc-4edf-8f2e-daef04ccc601"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.809381 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lfsh\" (UniqueName: \"kubernetes.io/projected/d915f7d2-5b4d-4017-a839-b615a182fafb-kube-api-access-8lfsh\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.809417 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75f42e15-d1dc-4edf-8f2e-daef04ccc601-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.844703 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:39 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:39 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:39 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:39 crc kubenswrapper[4733]: I0318 10:16:39.844769 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.063525 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"75f42e15-d1dc-4edf-8f2e-daef04ccc601","Type":"ContainerDied","Data":"d4949220d6469d09f06a115f37de0df57649271ca132ce8e25268effeb41c8de"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.063601 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4949220d6469d09f06a115f37de0df57649271ca132ce8e25268effeb41c8de" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.064057 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.068078 4733 generic.go:334] "Generic (PLEG): container finished" podID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerID="aa0522bdc088c10a6b3c5dba1e3ad5057a62e8ded941287c75083cef63e55041" exitCode=0 Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.068162 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerDied","Data":"aa0522bdc088c10a6b3c5dba1e3ad5057a62e8ded941287c75083cef63e55041"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.071840 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" event={"ID":"24e63f18-ff07-40b0-8289-351352d47d0a","Type":"ContainerStarted","Data":"9eef5c0d799a7506ebb7268521b337cf05cff4092c361a61789da71fa3bb245c"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.076330 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.096388 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.097213 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerStarted","Data":"1a1211028e93b8b114b76fa499d9200418412506c6795f17a8a464f56e421c4c"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.113993 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" event={"ID":"d915f7d2-5b4d-4017-a839-b615a182fafb","Type":"ContainerDied","Data":"4abe99dbd7bc3b694bde422289cb5e8d4d69c342990c6d29b9ffcb65e8f885f7"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.114067 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4abe99dbd7bc3b694bde422289cb5e8d4d69c342990c6d29b9ffcb65e8f885f7" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.114108 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.131575 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" podStartSLOduration=7.131558511 podStartE2EDuration="7.131558511s" podCreationTimestamp="2026-03-18 10:16:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:16:40.131178899 +0000 UTC m=+239.622913224" watchObservedRunningTime="2026-03-18 10:16:40.131558511 +0000 UTC m=+239.623292826" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.135374 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6503bd00-b300-438d-b10e-27380eaf7d9a","Type":"ContainerStarted","Data":"5097184223deb67237547660bb54a8b2f0c30328be9b17c9a3cc7d81a39aded5"} Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.249260 4733 ???:1] "http: TLS handshake error from 192.168.126.11:33552: no serving certificate available for the kubelet" Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.856182 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:40 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:40 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:40 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:40 crc kubenswrapper[4733]: I0318 10:16:40.856361 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.211785 4733 generic.go:334] "Generic (PLEG): container finished" podID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerID="9d9d502e889f0bc1ff5ac5bd25eb5937fb15878b89bb5f2186b3e420cda96e62" exitCode=0 Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.226013 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerDied","Data":"9d9d502e889f0bc1ff5ac5bd25eb5937fb15878b89bb5f2186b3e420cda96e62"} Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.241733 4733 generic.go:334] "Generic (PLEG): container finished" podID="6503bd00-b300-438d-b10e-27380eaf7d9a" containerID="5097184223deb67237547660bb54a8b2f0c30328be9b17c9a3cc7d81a39aded5" exitCode=0 Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.241897 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6503bd00-b300-438d-b10e-27380eaf7d9a","Type":"ContainerDied","Data":"5097184223deb67237547660bb54a8b2f0c30328be9b17c9a3cc7d81a39aded5"} Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.841575 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 10:16:41 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Mar 18 10:16:41 crc kubenswrapper[4733]: [+]process-running ok Mar 18 10:16:41 crc kubenswrapper[4733]: healthz check failed Mar 18 10:16:41 crc kubenswrapper[4733]: I0318 10:16:41.841631 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 10:16:42 crc kubenswrapper[4733]: I0318 10:16:42.796624 4733 ???:1] "http: TLS handshake error from 192.168.126.11:33562: no serving certificate available for the kubelet" Mar 18 10:16:42 crc kubenswrapper[4733]: I0318 10:16:42.844339 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:42 crc kubenswrapper[4733]: I0318 10:16:42.847302 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-xl5d7" Mar 18 10:16:42 crc kubenswrapper[4733]: I0318 10:16:42.898640 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hvmrz" Mar 18 10:16:43 crc kubenswrapper[4733]: I0318 10:16:43.571906 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:16:43 crc kubenswrapper[4733]: I0318 10:16:43.572001 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:16:43 crc kubenswrapper[4733]: I0318 10:16:43.605993 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:16:46 crc kubenswrapper[4733]: I0318 10:16:46.829780 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-gxcb2" Mar 18 10:16:46 crc kubenswrapper[4733]: I0318 10:16:46.922605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:46 crc kubenswrapper[4733]: I0318 10:16:46.928158 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:16:50 crc kubenswrapper[4733]: I0318 10:16:50.521807 4733 ???:1] "http: TLS handshake error from 192.168.126.11:42986: no serving certificate available for the kubelet" Mar 18 10:16:53 crc kubenswrapper[4733]: I0318 10:16:53.140102 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:53 crc kubenswrapper[4733]: I0318 10:16:53.140696 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" containerName="controller-manager" containerID="cri-o://9eef5c0d799a7506ebb7268521b337cf05cff4092c361a61789da71fa3bb245c" gracePeriod=30 Mar 18 10:16:53 crc kubenswrapper[4733]: I0318 10:16:53.165356 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:53 crc kubenswrapper[4733]: I0318 10:16:53.165645 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerName="route-controller-manager" containerID="cri-o://bb3b81003e39ecbea213f0c0b02b7e8dae8e9507c2c01e812e0dce4d4f5c71d4" gracePeriod=30 Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.202792 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.354995 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6503bd00-b300-438d-b10e-27380eaf7d9a","Type":"ContainerDied","Data":"cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e"} Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.355077 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cde036bce187894fcff4e3f9314bf8376952111eb9bf66cb5384b076ad589e4e" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.355070 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.357053 4733 generic.go:334] "Generic (PLEG): container finished" podID="24e63f18-ff07-40b0-8289-351352d47d0a" containerID="9eef5c0d799a7506ebb7268521b337cf05cff4092c361a61789da71fa3bb245c" exitCode=0 Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.357179 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" event={"ID":"24e63f18-ff07-40b0-8289-351352d47d0a","Type":"ContainerDied","Data":"9eef5c0d799a7506ebb7268521b337cf05cff4092c361a61789da71fa3bb245c"} Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.361955 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerID="bb3b81003e39ecbea213f0c0b02b7e8dae8e9507c2c01e812e0dce4d4f5c71d4" exitCode=0 Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.362218 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" event={"ID":"d1954584-f4d6-467a-8b0c-9db32f9e385c","Type":"ContainerDied","Data":"bb3b81003e39ecbea213f0c0b02b7e8dae8e9507c2c01e812e0dce4d4f5c71d4"} Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.384690 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir\") pod \"6503bd00-b300-438d-b10e-27380eaf7d9a\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.384817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access\") pod \"6503bd00-b300-438d-b10e-27380eaf7d9a\" (UID: \"6503bd00-b300-438d-b10e-27380eaf7d9a\") " Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.384855 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6503bd00-b300-438d-b10e-27380eaf7d9a" (UID: "6503bd00-b300-438d-b10e-27380eaf7d9a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.385906 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6503bd00-b300-438d-b10e-27380eaf7d9a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.410473 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6503bd00-b300-438d-b10e-27380eaf7d9a" (UID: "6503bd00-b300-438d-b10e-27380eaf7d9a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:54 crc kubenswrapper[4733]: I0318 10:16:54.486804 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6503bd00-b300-438d-b10e-27380eaf7d9a-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:55 crc kubenswrapper[4733]: I0318 10:16:55.728519 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:16:55 crc kubenswrapper[4733]: I0318 10:16:55.972393 4733 patch_prober.go:28] interesting pod/route-controller-manager-6946475f8-lnppg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Mar 18 10:16:55 crc kubenswrapper[4733]: I0318 10:16:55.972437 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.471285 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.471738 4733 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 10:16:57 crc kubenswrapper[4733]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 18 10:16:57 crc kubenswrapper[4733]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2whnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29563816-4582s_openshift-infra(71a70c3c-d483-43f4-9f54-10978c7f8cc8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 18 10:16:57 crc kubenswrapper[4733]: > logger="UnhandledError" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.472832 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29563816-4582s" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.807757 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.820500 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867210 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.867513 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" containerName="controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867526 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" containerName="controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.867545 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6503bd00-b300-438d-b10e-27380eaf7d9a" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867552 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6503bd00-b300-438d-b10e-27380eaf7d9a" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.867565 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerName="route-controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867574 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerName="route-controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.867585 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d915f7d2-5b4d-4017-a839-b615a182fafb" containerName="collect-profiles" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867591 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d915f7d2-5b4d-4017-a839-b615a182fafb" containerName="collect-profiles" Mar 18 10:16:57 crc kubenswrapper[4733]: E0318 10:16:57.867598 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f42e15-d1dc-4edf-8f2e-daef04ccc601" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867604 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f42e15-d1dc-4edf-8f2e-daef04ccc601" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867732 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d915f7d2-5b4d-4017-a839-b615a182fafb" containerName="collect-profiles" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867742 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f42e15-d1dc-4edf-8f2e-daef04ccc601" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867753 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" containerName="route-controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867762 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" containerName="controller-manager" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.867772 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6503bd00-b300-438d-b10e-27380eaf7d9a" containerName="pruner" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.868139 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.869740 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934127 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert\") pod \"d1954584-f4d6-467a-8b0c-9db32f9e385c\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934205 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert\") pod \"24e63f18-ff07-40b0-8289-351352d47d0a\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934247 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmg68\" (UniqueName: \"kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68\") pod \"d1954584-f4d6-467a-8b0c-9db32f9e385c\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934306 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config\") pod \"24e63f18-ff07-40b0-8289-351352d47d0a\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934367 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config\") pod \"d1954584-f4d6-467a-8b0c-9db32f9e385c\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934391 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca\") pod \"d1954584-f4d6-467a-8b0c-9db32f9e385c\" (UID: \"d1954584-f4d6-467a-8b0c-9db32f9e385c\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.934420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvgcj\" (UniqueName: \"kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj\") pod \"24e63f18-ff07-40b0-8289-351352d47d0a\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles\") pod \"24e63f18-ff07-40b0-8289-351352d47d0a\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935347 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config" (OuterVolumeSpecName: "config") pod "d1954584-f4d6-467a-8b0c-9db32f9e385c" (UID: "d1954584-f4d6-467a-8b0c-9db32f9e385c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935403 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca\") pod \"24e63f18-ff07-40b0-8289-351352d47d0a\" (UID: \"24e63f18-ff07-40b0-8289-351352d47d0a\") " Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935451 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config" (OuterVolumeSpecName: "config") pod "24e63f18-ff07-40b0-8289-351352d47d0a" (UID: "24e63f18-ff07-40b0-8289-351352d47d0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935449 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca" (OuterVolumeSpecName: "client-ca") pod "d1954584-f4d6-467a-8b0c-9db32f9e385c" (UID: "d1954584-f4d6-467a-8b0c-9db32f9e385c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.935865 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "24e63f18-ff07-40b0-8289-351352d47d0a" (UID: "24e63f18-ff07-40b0-8289-351352d47d0a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.936729 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca" (OuterVolumeSpecName: "client-ca") pod "24e63f18-ff07-40b0-8289-351352d47d0a" (UID: "24e63f18-ff07-40b0-8289-351352d47d0a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.937060 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.937087 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.937099 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1954584-f4d6-467a-8b0c-9db32f9e385c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.937110 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.937125 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24e63f18-ff07-40b0-8289-351352d47d0a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.940692 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj" (OuterVolumeSpecName: "kube-api-access-nvgcj") pod "24e63f18-ff07-40b0-8289-351352d47d0a" (UID: "24e63f18-ff07-40b0-8289-351352d47d0a"). InnerVolumeSpecName "kube-api-access-nvgcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.941804 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68" (OuterVolumeSpecName: "kube-api-access-dmg68") pod "d1954584-f4d6-467a-8b0c-9db32f9e385c" (UID: "d1954584-f4d6-467a-8b0c-9db32f9e385c"). InnerVolumeSpecName "kube-api-access-dmg68". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.943111 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "24e63f18-ff07-40b0-8289-351352d47d0a" (UID: "24e63f18-ff07-40b0-8289-351352d47d0a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:16:57 crc kubenswrapper[4733]: I0318 10:16:57.952719 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d1954584-f4d6-467a-8b0c-9db32f9e385c" (UID: "d1954584-f4d6-467a-8b0c-9db32f9e385c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.037948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.037999 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzm9t\" (UniqueName: \"kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038342 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038403 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038545 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvgcj\" (UniqueName: \"kubernetes.io/projected/24e63f18-ff07-40b0-8289-351352d47d0a-kube-api-access-nvgcj\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038596 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1954584-f4d6-467a-8b0c-9db32f9e385c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038610 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24e63f18-ff07-40b0-8289-351352d47d0a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.038623 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmg68\" (UniqueName: \"kubernetes.io/projected/d1954584-f4d6-467a-8b0c-9db32f9e385c-kube-api-access-dmg68\") on node \"crc\" DevicePath \"\"" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.139405 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.139460 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.139516 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.139535 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.139557 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzm9t\" (UniqueName: \"kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.140983 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.141396 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.141770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.145779 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.157988 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzm9t\" (UniqueName: \"kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t\") pod \"controller-manager-6748588445-r4bnh\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.191399 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.391250 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.391293 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" event={"ID":"24e63f18-ff07-40b0-8289-351352d47d0a","Type":"ContainerDied","Data":"4e6f1dbd850522a9734d4ded476f261e3ae7fcfb8b06688f8b915837852ee325"} Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.391698 4733 scope.go:117] "RemoveContainer" containerID="9eef5c0d799a7506ebb7268521b337cf05cff4092c361a61789da71fa3bb245c" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.394028 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.394050 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg" event={"ID":"d1954584-f4d6-467a-8b0c-9db32f9e385c","Type":"ContainerDied","Data":"f9b9ef0617ed480f3b6667ef58bad6aa0295e3670f7f52f36e961678f655af59"} Mar 18 10:16:58 crc kubenswrapper[4733]: E0318 10:16:58.395535 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29563816-4582s" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.424936 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.430144 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-75c9fbd49-qmfds"] Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.439449 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.442364 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6946475f8-lnppg"] Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.741038 4733 patch_prober.go:28] interesting pod/controller-manager-75c9fbd49-qmfds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:16:58 crc kubenswrapper[4733]: I0318 10:16:58.741101 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-75c9fbd49-qmfds" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:16:59 crc kubenswrapper[4733]: I0318 10:16:59.184518 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24e63f18-ff07-40b0-8289-351352d47d0a" path="/var/lib/kubelet/pods/24e63f18-ff07-40b0-8289-351352d47d0a/volumes" Mar 18 10:16:59 crc kubenswrapper[4733]: I0318 10:16:59.185651 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1954584-f4d6-467a-8b0c-9db32f9e385c" path="/var/lib/kubelet/pods/d1954584-f4d6-467a-8b0c-9db32f9e385c/volumes" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.428971 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.430323 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.432989 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.433079 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.433161 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.434680 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.441299 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.441576 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.446724 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.607482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.607619 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzjn9\" (UniqueName: \"kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.607700 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.607736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.709805 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.709893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzjn9\" (UniqueName: \"kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.709942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.709961 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.711343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.711461 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.717139 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.725274 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzjn9\" (UniqueName: \"kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9\") pod \"route-controller-manager-6d8fb44d7f-zrsx6\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:02 crc kubenswrapper[4733]: I0318 10:17:02.778485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:07 crc kubenswrapper[4733]: I0318 10:17:07.859143 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" Mar 18 10:17:09 crc kubenswrapper[4733]: E0318 10:17:09.754049 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 18 10:17:09 crc kubenswrapper[4733]: E0318 10:17:09.754311 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5vpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w7rrs_openshift-marketplace(02cd6358-355c-4db8-b0f7-2528618602ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 10:17:09 crc kubenswrapper[4733]: E0318 10:17:09.755989 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-w7rrs" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" Mar 18 10:17:11 crc kubenswrapper[4733]: E0318 10:17:11.359396 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w7rrs" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" Mar 18 10:17:11 crc kubenswrapper[4733]: I0318 10:17:11.897635 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 10:17:11 crc kubenswrapper[4733]: I0318 10:17:11.898403 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:11 crc kubenswrapper[4733]: I0318 10:17:11.900670 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 10:17:11 crc kubenswrapper[4733]: I0318 10:17:11.900998 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 10:17:11 crc kubenswrapper[4733]: I0318 10:17:11.915782 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.090017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.090133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.157245 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.157623 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5frtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jb86w_openshift-marketplace(0fd306cb-05db-40e1-a1ec-9f811ce7fec0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.159036 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jb86w" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.191691 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.191832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.192265 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.214123 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.262123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:12 crc kubenswrapper[4733]: I0318 10:17:12.330927 4733 scope.go:117] "RemoveContainer" containerID="bb3b81003e39ecbea213f0c0b02b7e8dae8e9507c2c01e812e0dce4d4f5c71d4" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.916211 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.916890 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7hv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rls2r_openshift-marketplace(92996997-080b-42c9-bc2c-19c2e68db896): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 10:17:12 crc kubenswrapper[4733]: E0318 10:17:12.918572 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rls2r" podUID="92996997-080b-42c9-bc2c-19c2e68db896" Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.122965 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jb86w" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" Mar 18 10:17:13 crc kubenswrapper[4733]: I0318 10:17:13.141115 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:17:13 crc kubenswrapper[4733]: I0318 10:17:13.249951 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.295275 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.295572 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6r2vl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-f82xf_openshift-marketplace(c91f12fa-96f0-442a-a3f7-70d56a697839): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.296971 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-f82xf" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" Mar 18 10:17:13 crc kubenswrapper[4733]: I0318 10:17:13.571683 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:17:13 crc kubenswrapper[4733]: I0318 10:17:13.571756 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.641709 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rls2r" podUID="92996997-080b-42c9-bc2c-19c2e68db896" Mar 18 10:17:13 crc kubenswrapper[4733]: E0318 10:17:13.641944 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-f82xf" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.193643 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:17:14 crc kubenswrapper[4733]: W0318 10:17:14.204592 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19d4bb8b_0ef7_4aae_9eca_d24dbb957f8b.slice/crio-729502b90d95e5266fca1fcfc3473b605230425c19cf5542d40820061af70fa1 WatchSource:0}: Error finding container 729502b90d95e5266fca1fcfc3473b605230425c19cf5542d40820061af70fa1: Status 404 returned error can't find the container with id 729502b90d95e5266fca1fcfc3473b605230425c19cf5542d40820061af70fa1 Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.279729 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.286983 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.482423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" event={"ID":"8ad622ed-fcbc-4485-a927-639a658660fa","Type":"ContainerStarted","Data":"c17a0094029923b923cf507d45924db60f811cb6ad73561bc864ff132db3d533"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.483537 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"068ddaf0-0f79-459b-b064-1b90505c36ca","Type":"ContainerStarted","Data":"a37d483c7ee4ef126cb1f876ea9f88989d5b67cbbc348139e49a23862aa79e05"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.499926 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerStarted","Data":"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.505095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerStarted","Data":"340cfa7d2b8654b1dea28355651bf6f54381a8104d827e6d38142ffcaf93e8ae"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.513115 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" event={"ID":"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b","Type":"ContainerStarted","Data":"729502b90d95e5266fca1fcfc3473b605230425c19cf5542d40820061af70fa1"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.516133 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerStarted","Data":"29a9561b8927709b5dd59a92cbf81b78eacad45dd6ac5ec49191d6faee246d53"} Mar 18 10:17:14 crc kubenswrapper[4733]: I0318 10:17:14.519343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerStarted","Data":"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.530669 4733 generic.go:334] "Generic (PLEG): container finished" podID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerID="29a9561b8927709b5dd59a92cbf81b78eacad45dd6ac5ec49191d6faee246d53" exitCode=0 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.530758 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerDied","Data":"29a9561b8927709b5dd59a92cbf81b78eacad45dd6ac5ec49191d6faee246d53"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.533409 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563816-4582s" event={"ID":"71a70c3c-d483-43f4-9f54-10978c7f8cc8","Type":"ContainerStarted","Data":"869578488a5526adb52c0d5efeb676ea68e5c20e95b1cf2d208fa00dbd02baca"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.536488 4733 generic.go:334] "Generic (PLEG): container finished" podID="527056ad-4daf-4dd5-9e31-887d55be0336" containerID="29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4" exitCode=0 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.536607 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerDied","Data":"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.538624 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" event={"ID":"8ad622ed-fcbc-4485-a927-639a658660fa","Type":"ContainerStarted","Data":"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.538744 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" podUID="8ad622ed-fcbc-4485-a927-639a658660fa" containerName="route-controller-manager" containerID="cri-o://363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b" gracePeriod=30 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.538906 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.542667 4733 generic.go:334] "Generic (PLEG): container finished" podID="068ddaf0-0f79-459b-b064-1b90505c36ca" containerID="88cfa7738cd509c2ad7f6041d284fa78af0a64d896c50f75f0a19f578d7cb91d" exitCode=0 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.542789 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"068ddaf0-0f79-459b-b064-1b90505c36ca","Type":"ContainerDied","Data":"88cfa7738cd509c2ad7f6041d284fa78af0a64d896c50f75f0a19f578d7cb91d"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.547080 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.547596 4733 generic.go:334] "Generic (PLEG): container finished" podID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerID="c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c" exitCode=0 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.547640 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerDied","Data":"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.551313 4733 generic.go:334] "Generic (PLEG): container finished" podID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerID="340cfa7d2b8654b1dea28355651bf6f54381a8104d827e6d38142ffcaf93e8ae" exitCode=0 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.551401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerDied","Data":"340cfa7d2b8654b1dea28355651bf6f54381a8104d827e6d38142ffcaf93e8ae"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.553054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" event={"ID":"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b","Type":"ContainerStarted","Data":"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277"} Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.553179 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" podUID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" containerName="controller-manager" containerID="cri-o://99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277" gracePeriod=30 Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.553308 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.562322 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.583475 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" podStartSLOduration=22.583454564 podStartE2EDuration="22.583454564s" podCreationTimestamp="2026-03-18 10:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:15.581768987 +0000 UTC m=+275.073503312" watchObservedRunningTime="2026-03-18 10:17:15.583454564 +0000 UTC m=+275.075188889" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.634636 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" podStartSLOduration=22.634603006 podStartE2EDuration="22.634603006s" podCreationTimestamp="2026-03-18 10:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:15.632468076 +0000 UTC m=+275.124202401" watchObservedRunningTime="2026-03-18 10:17:15.634603006 +0000 UTC m=+275.126337331" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.684222 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563816-4582s" podStartSLOduration=29.799564362 podStartE2EDuration="1m15.684167824s" podCreationTimestamp="2026-03-18 10:16:00 +0000 UTC" firstStartedPulling="2026-03-18 10:16:28.967450925 +0000 UTC m=+228.459185250" lastFinishedPulling="2026-03-18 10:17:14.852054387 +0000 UTC m=+274.343788712" observedRunningTime="2026-03-18 10:17:15.674881114 +0000 UTC m=+275.166615439" watchObservedRunningTime="2026-03-18 10:17:15.684167824 +0000 UTC m=+275.175902149" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.694352 4733 csr.go:261] certificate signing request csr-8nwhj is approved, waiting to be issued Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.711933 4733 csr.go:257] certificate signing request csr-8nwhj is issued Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.973656 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:15 crc kubenswrapper[4733]: I0318 10:17:15.982551 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.004110 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:16 crc kubenswrapper[4733]: E0318 10:17:16.004933 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad622ed-fcbc-4485-a927-639a658660fa" containerName="route-controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.004953 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad622ed-fcbc-4485-a927-639a658660fa" containerName="route-controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: E0318 10:17:16.004972 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" containerName="controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.004979 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" containerName="controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.005076 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" containerName="controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.005096 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad622ed-fcbc-4485-a927-639a658660fa" containerName="route-controller-manager" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.005489 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.016183 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.051015 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.051220 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.051283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.051301 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwfk4\" (UniqueName: \"kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.152847 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzm9t\" (UniqueName: \"kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t\") pod \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.152918 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config\") pod \"8ad622ed-fcbc-4485-a927-639a658660fa\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153015 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert\") pod \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153236 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca\") pod \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153292 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config\") pod \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153332 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzjn9\" (UniqueName: \"kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9\") pod \"8ad622ed-fcbc-4485-a927-639a658660fa\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153398 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles\") pod \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\" (UID: \"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153454 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert\") pod \"8ad622ed-fcbc-4485-a927-639a658660fa\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca\") pod \"8ad622ed-fcbc-4485-a927-639a658660fa\" (UID: \"8ad622ed-fcbc-4485-a927-639a658660fa\") " Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153796 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153859 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153905 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.153927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwfk4\" (UniqueName: \"kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.154102 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca" (OuterVolumeSpecName: "client-ca") pod "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" (UID: "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.155024 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config" (OuterVolumeSpecName: "config") pod "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" (UID: "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.155354 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca" (OuterVolumeSpecName: "client-ca") pod "8ad622ed-fcbc-4485-a927-639a658660fa" (UID: "8ad622ed-fcbc-4485-a927-639a658660fa"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.155629 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" (UID: "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.157004 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.157229 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.157223 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config" (OuterVolumeSpecName: "config") pod "8ad622ed-fcbc-4485-a927-639a658660fa" (UID: "8ad622ed-fcbc-4485-a927-639a658660fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.159736 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" (UID: "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.160276 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8ad622ed-fcbc-4485-a927-639a658660fa" (UID: "8ad622ed-fcbc-4485-a927-639a658660fa"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.160276 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9" (OuterVolumeSpecName: "kube-api-access-pzjn9") pod "8ad622ed-fcbc-4485-a927-639a658660fa" (UID: "8ad622ed-fcbc-4485-a927-639a658660fa"). InnerVolumeSpecName "kube-api-access-pzjn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.160374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t" (OuterVolumeSpecName: "kube-api-access-hzm9t") pod "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" (UID: "19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b"). InnerVolumeSpecName "kube-api-access-hzm9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.160554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.171886 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwfk4\" (UniqueName: \"kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4\") pod \"route-controller-manager-6966b4866d-xm6hn\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255737 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255785 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ad622ed-fcbc-4485-a927-639a658660fa-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255798 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255809 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzm9t\" (UniqueName: \"kubernetes.io/projected/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-kube-api-access-hzm9t\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255824 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad622ed-fcbc-4485-a927-639a658660fa-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255837 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255849 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255860 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.255873 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzjn9\" (UniqueName: \"kubernetes.io/projected/8ad622ed-fcbc-4485-a927-639a658660fa-kube-api-access-pzjn9\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.406530 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.572296 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerStarted","Data":"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.575888 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerStarted","Data":"33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.577516 4733 generic.go:334] "Generic (PLEG): container finished" podID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" containerID="99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277" exitCode=0 Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.577590 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" event={"ID":"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b","Type":"ContainerDied","Data":"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.577636 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" event={"ID":"19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b","Type":"ContainerDied","Data":"729502b90d95e5266fca1fcfc3473b605230425c19cf5542d40820061af70fa1"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.577636 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6748588445-r4bnh" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.577663 4733 scope.go:117] "RemoveContainer" containerID="99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.591316 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerStarted","Data":"7a59787487ae6ad4c8775bad7fe0e44e006e25d60fd069b4d5bd8cc6ceca6c70"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.603717 4733 generic.go:334] "Generic (PLEG): container finished" podID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" containerID="869578488a5526adb52c0d5efeb676ea68e5c20e95b1cf2d208fa00dbd02baca" exitCode=0 Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.604222 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563816-4582s" event={"ID":"71a70c3c-d483-43f4-9f54-10978c7f8cc8","Type":"ContainerDied","Data":"869578488a5526adb52c0d5efeb676ea68e5c20e95b1cf2d208fa00dbd02baca"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.616702 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gmw2d" podStartSLOduration=3.414821918 podStartE2EDuration="42.616668181s" podCreationTimestamp="2026-03-18 10:16:34 +0000 UTC" firstStartedPulling="2026-03-18 10:16:36.800759136 +0000 UTC m=+236.292493471" lastFinishedPulling="2026-03-18 10:17:16.002605409 +0000 UTC m=+275.494339734" observedRunningTime="2026-03-18 10:17:16.613763599 +0000 UTC m=+276.105497934" watchObservedRunningTime="2026-03-18 10:17:16.616668181 +0000 UTC m=+276.108402506" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.628171 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerStarted","Data":"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.640002 4733 scope.go:117] "RemoveContainer" containerID="99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277" Mar 18 10:17:16 crc kubenswrapper[4733]: E0318 10:17:16.640799 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277\": container with ID starting with 99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277 not found: ID does not exist" containerID="99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.640882 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277"} err="failed to get container status \"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277\": rpc error: code = NotFound desc = could not find container \"99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277\": container with ID starting with 99d80b19731c17ac26d38096f7c5b3a305c502e2750b077f48abc2058aa5d277 not found: ID does not exist" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.640922 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ad622ed-fcbc-4485-a927-639a658660fa" containerID="363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b" exitCode=0 Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.640996 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.640992 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" event={"ID":"8ad622ed-fcbc-4485-a927-639a658660fa","Type":"ContainerDied","Data":"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.641053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6" event={"ID":"8ad622ed-fcbc-4485-a927-639a658660fa","Type":"ContainerDied","Data":"c17a0094029923b923cf507d45924db60f811cb6ad73561bc864ff132db3d533"} Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.641078 4733 scope.go:117] "RemoveContainer" containerID="363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.656742 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrwxg" podStartSLOduration=3.497819402 podStartE2EDuration="39.656727892s" podCreationTimestamp="2026-03-18 10:16:37 +0000 UTC" firstStartedPulling="2026-03-18 10:16:40.070658083 +0000 UTC m=+239.562392408" lastFinishedPulling="2026-03-18 10:17:16.229566573 +0000 UTC m=+275.721300898" observedRunningTime="2026-03-18 10:17:16.655544269 +0000 UTC m=+276.147278584" watchObservedRunningTime="2026-03-18 10:17:16.656727892 +0000 UTC m=+276.148462217" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.669395 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.678222 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6748588445-r4bnh"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.689176 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ll7dp" podStartSLOduration=15.974903489 podStartE2EDuration="39.68916051s" podCreationTimestamp="2026-03-18 10:16:37 +0000 UTC" firstStartedPulling="2026-03-18 10:16:52.239528192 +0000 UTC m=+251.731262537" lastFinishedPulling="2026-03-18 10:17:15.953785233 +0000 UTC m=+275.445519558" observedRunningTime="2026-03-18 10:17:16.68699282 +0000 UTC m=+276.178727155" watchObservedRunningTime="2026-03-18 10:17:16.68916051 +0000 UTC m=+276.180894825" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.703481 4733 scope.go:117] "RemoveContainer" containerID="363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b" Mar 18 10:17:16 crc kubenswrapper[4733]: E0318 10:17:16.707793 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b\": container with ID starting with 363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b not found: ID does not exist" containerID="363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.707831 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b"} err="failed to get container status \"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b\": rpc error: code = NotFound desc = could not find container \"363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b\": container with ID starting with 363048c696b7c0c6504ff378d37d2a0ff6eece67133ad5d68c872cadbe32058b not found: ID does not exist" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.717341 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-13 13:56:53.35255902 +0000 UTC Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.717405 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5763h39m36.635157219s for next certificate rotation Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.732133 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.750672 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f92nl" podStartSLOduration=2.622889314 podStartE2EDuration="42.750646152s" podCreationTimestamp="2026-03-18 10:16:34 +0000 UTC" firstStartedPulling="2026-03-18 10:16:35.812810954 +0000 UTC m=+235.304545279" lastFinishedPulling="2026-03-18 10:17:15.940567792 +0000 UTC m=+275.432302117" observedRunningTime="2026-03-18 10:17:16.734540251 +0000 UTC m=+276.226274576" watchObservedRunningTime="2026-03-18 10:17:16.750646152 +0000 UTC m=+276.242380477" Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.761081 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.777479 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d8fb44d7f-zrsx6"] Mar 18 10:17:16 crc kubenswrapper[4733]: I0318 10:17:16.988698 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.169647 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir\") pod \"068ddaf0-0f79-459b-b064-1b90505c36ca\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.169739 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access\") pod \"068ddaf0-0f79-459b-b064-1b90505c36ca\" (UID: \"068ddaf0-0f79-459b-b064-1b90505c36ca\") " Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.169800 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "068ddaf0-0f79-459b-b064-1b90505c36ca" (UID: "068ddaf0-0f79-459b-b064-1b90505c36ca"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.170043 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/068ddaf0-0f79-459b-b064-1b90505c36ca-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.180486 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "068ddaf0-0f79-459b-b064-1b90505c36ca" (UID: "068ddaf0-0f79-459b-b064-1b90505c36ca"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.195833 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b" path="/var/lib/kubelet/pods/19d4bb8b-0ef7-4aae-9eca-d24dbb957f8b/volumes" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.197024 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad622ed-fcbc-4485-a927-639a658660fa" path="/var/lib/kubelet/pods/8ad622ed-fcbc-4485-a927-639a658660fa/volumes" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.271433 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/068ddaf0-0f79-459b-b064-1b90505c36ca-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.639297 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.641175 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.650404 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" event={"ID":"debb4f43-7f9a-4fdc-9896-db5106650a74","Type":"ContainerStarted","Data":"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740"} Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.650431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" event={"ID":"debb4f43-7f9a-4fdc-9896-db5106650a74","Type":"ContainerStarted","Data":"f30c8eb187bc3102906d7afcd8a11a3b91672d508cc51863174f4d3b52b10205"} Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.651470 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.656241 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"068ddaf0-0f79-459b-b064-1b90505c36ca","Type":"ContainerDied","Data":"a37d483c7ee4ef126cb1f876ea9f88989d5b67cbbc348139e49a23862aa79e05"} Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.656311 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a37d483c7ee4ef126cb1f876ea9f88989d5b67cbbc348139e49a23862aa79e05" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.656350 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.667607 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.676428 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" podStartSLOduration=4.676403871 podStartE2EDuration="4.676403871s" podCreationTimestamp="2026-03-18 10:17:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:17.674357064 +0000 UTC m=+277.166091399" watchObservedRunningTime="2026-03-18 10:17:17.676403871 +0000 UTC m=+277.168138206" Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.718092 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-06 05:20:34.513480736 +0000 UTC Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.718535 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6307h3m16.795158741s for next certificate rotation Mar 18 10:17:17 crc kubenswrapper[4733]: I0318 10:17:17.957331 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.087975 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2whnv\" (UniqueName: \"kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv\") pod \"71a70c3c-d483-43f4-9f54-10978c7f8cc8\" (UID: \"71a70c3c-d483-43f4-9f54-10978c7f8cc8\") " Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.098832 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv" (OuterVolumeSpecName: "kube-api-access-2whnv") pod "71a70c3c-d483-43f4-9f54-10978c7f8cc8" (UID: "71a70c3c-d483-43f4-9f54-10978c7f8cc8"). InnerVolumeSpecName "kube-api-access-2whnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.189231 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2whnv\" (UniqueName: \"kubernetes.io/projected/71a70c3c-d483-43f4-9f54-10978c7f8cc8-kube-api-access-2whnv\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.337848 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.337927 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.446259 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:18 crc kubenswrapper[4733]: E0318 10:17:18.446720 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" containerName="oc" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.446733 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" containerName="oc" Mar 18 10:17:18 crc kubenswrapper[4733]: E0318 10:17:18.446748 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068ddaf0-0f79-459b-b064-1b90505c36ca" containerName="pruner" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.446755 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="068ddaf0-0f79-459b-b064-1b90505c36ca" containerName="pruner" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.446853 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" containerName="oc" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.446864 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="068ddaf0-0f79-459b-b064-1b90505c36ca" containerName="pruner" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.447213 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.448635 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.452218 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.452784 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.452841 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.453069 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.454476 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.469842 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.470019 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.528682 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.528751 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.528817 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.528869 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575kc\" (UniqueName: \"kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.528898 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.630213 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.630313 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575kc\" (UniqueName: \"kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.630343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.630408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.630439 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.631647 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.631966 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.632110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.638958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.656794 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575kc\" (UniqueName: \"kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc\") pod \"controller-manager-7cf89766c-wb795\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.670478 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563816-4582s" event={"ID":"71a70c3c-d483-43f4-9f54-10978c7f8cc8","Type":"ContainerDied","Data":"f9019fd1aca4002d61050c62413d5f0b6ff4613e81da7416fe1c8a2924a20e45"} Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.670527 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9019fd1aca4002d61050c62413d5f0b6ff4613e81da7416fe1c8a2924a20e45" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.670769 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563816-4582s" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.759592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:18 crc kubenswrapper[4733]: I0318 10:17:18.818604 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrwxg" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" probeResult="failure" output=< Mar 18 10:17:18 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:17:18 crc kubenswrapper[4733]: > Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.006710 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:19 crc kubenswrapper[4733]: W0318 10:17:19.013495 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf280b8dd_9b2f_4faa_9335_ca84961ea0a6.slice/crio-15b644538c3645c79d06f703952a07b7b4ae39d43d1dc32a31009b361d9095cd WatchSource:0}: Error finding container 15b644538c3645c79d06f703952a07b7b4ae39d43d1dc32a31009b361d9095cd: Status 404 returned error can't find the container with id 15b644538c3645c79d06f703952a07b7b4ae39d43d1dc32a31009b361d9095cd Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.385870 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ll7dp" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="registry-server" probeResult="failure" output=< Mar 18 10:17:19 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:17:19 crc kubenswrapper[4733]: > Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.683662 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" event={"ID":"f280b8dd-9b2f-4faa-9335-ca84961ea0a6","Type":"ContainerStarted","Data":"c282a4dca9ca3d5260c23c43faf22ce611247dc38c8752227edd1fbf81bf90dd"} Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.683712 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" event={"ID":"f280b8dd-9b2f-4faa-9335-ca84961ea0a6","Type":"ContainerStarted","Data":"15b644538c3645c79d06f703952a07b7b4ae39d43d1dc32a31009b361d9095cd"} Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.684016 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.689060 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:19 crc kubenswrapper[4733]: I0318 10:17:19.701124 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" podStartSLOduration=6.7011000670000005 podStartE2EDuration="6.701100067s" podCreationTimestamp="2026-03-18 10:17:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:19.69834471 +0000 UTC m=+279.190079045" watchObservedRunningTime="2026-03-18 10:17:19.701100067 +0000 UTC m=+279.192834392" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.695800 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.696812 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.701902 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.702182 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.760806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.760911 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.760937 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.862110 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.862625 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.862696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.862793 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.862840 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:20 crc kubenswrapper[4733]: I0318 10:17:20.892063 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access\") pod \"installer-9-crc\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:21 crc kubenswrapper[4733]: I0318 10:17:21.022280 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:17:21 crc kubenswrapper[4733]: I0318 10:17:21.160720 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 10:17:21 crc kubenswrapper[4733]: I0318 10:17:21.413022 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 10:17:22 crc kubenswrapper[4733]: I0318 10:17:22.139111 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"71bc6618-8df4-4a35-9469-772a853eff06","Type":"ContainerStarted","Data":"5f44602afd79b72c25bccf945c72fd688dbc42ff6b86533bf0722398bd85fb3d"} Mar 18 10:17:23 crc kubenswrapper[4733]: I0318 10:17:23.144921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"71bc6618-8df4-4a35-9469-772a853eff06","Type":"ContainerStarted","Data":"7f5d2a4800b0b935a593d622bab229709f8902d75a6b9d3d310047bf50063a1a"} Mar 18 10:17:23 crc kubenswrapper[4733]: I0318 10:17:23.150758 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerStarted","Data":"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450"} Mar 18 10:17:23 crc kubenswrapper[4733]: I0318 10:17:23.189387 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.18934915 podStartE2EDuration="3.18934915s" podCreationTimestamp="2026-03-18 10:17:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:23.181056997 +0000 UTC m=+282.672791322" watchObservedRunningTime="2026-03-18 10:17:23.18934915 +0000 UTC m=+282.681083495" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.158674 4733 generic.go:334] "Generic (PLEG): container finished" podID="02cd6358-355c-4db8-b0f7-2528618602ff" containerID="6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450" exitCode=0 Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.158774 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerDied","Data":"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450"} Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.503411 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.503480 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.560416 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.859045 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.859453 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:24 crc kubenswrapper[4733]: I0318 10:17:24.900908 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:25 crc kubenswrapper[4733]: I0318 10:17:25.168034 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerStarted","Data":"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6"} Mar 18 10:17:25 crc kubenswrapper[4733]: I0318 10:17:25.199177 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w7rrs" podStartSLOduration=3.472350661 podStartE2EDuration="51.199154028s" podCreationTimestamp="2026-03-18 10:16:34 +0000 UTC" firstStartedPulling="2026-03-18 10:16:36.811562101 +0000 UTC m=+236.303296426" lastFinishedPulling="2026-03-18 10:17:24.538365448 +0000 UTC m=+284.030099793" observedRunningTime="2026-03-18 10:17:25.196106243 +0000 UTC m=+284.687840578" watchObservedRunningTime="2026-03-18 10:17:25.199154028 +0000 UTC m=+284.690888363" Mar 18 10:17:25 crc kubenswrapper[4733]: I0318 10:17:25.232544 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:25 crc kubenswrapper[4733]: I0318 10:17:25.234203 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:17:26 crc kubenswrapper[4733]: I0318 10:17:26.177922 4733 generic.go:334] "Generic (PLEG): container finished" podID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerID="4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9" exitCode=0 Mar 18 10:17:26 crc kubenswrapper[4733]: I0318 10:17:26.178082 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerDied","Data":"4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9"} Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.188949 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerStarted","Data":"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014"} Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.189924 4733 generic.go:334] "Generic (PLEG): container finished" podID="92996997-080b-42c9-bc2c-19c2e68db896" containerID="9295312051c24cc07301903e63a22c698207253e2dd4d338c0be4c6fd4de6dec" exitCode=0 Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.189971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerDied","Data":"9295312051c24cc07301903e63a22c698207253e2dd4d338c0be4c6fd4de6dec"} Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.251394 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jb86w" podStartSLOduration=3.614479629 podStartE2EDuration="52.251358894s" podCreationTimestamp="2026-03-18 10:16:35 +0000 UTC" firstStartedPulling="2026-03-18 10:16:37.924782901 +0000 UTC m=+237.416517226" lastFinishedPulling="2026-03-18 10:17:26.561662156 +0000 UTC m=+286.053396491" observedRunningTime="2026-03-18 10:17:27.24978807 +0000 UTC m=+286.741522425" watchObservedRunningTime="2026-03-18 10:17:27.251358894 +0000 UTC m=+286.743093229" Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.694777 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:17:27 crc kubenswrapper[4733]: I0318 10:17:27.748824 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:17:28 crc kubenswrapper[4733]: I0318 10:17:28.405702 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:28 crc kubenswrapper[4733]: I0318 10:17:28.460150 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:29 crc kubenswrapper[4733]: I0318 10:17:29.423180 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:17:29 crc kubenswrapper[4733]: I0318 10:17:29.423880 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gmw2d" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="registry-server" containerID="cri-o://59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d" gracePeriod=2 Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.075653 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.216156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content\") pod \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.216267 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs5rx\" (UniqueName: \"kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx\") pod \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.216376 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities\") pod \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\" (UID: \"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc\") " Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.217246 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities" (OuterVolumeSpecName: "utilities") pod "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" (UID: "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.219506 4733 generic.go:334] "Generic (PLEG): container finished" podID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerID="59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d" exitCode=0 Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.219580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerDied","Data":"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d"} Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.219615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmw2d" event={"ID":"7eb97f2d-18fa-4e8c-895f-de4602c9dbbc","Type":"ContainerDied","Data":"a2b42f75b17ecdce018f92ac6406accaeca335b14c1245cfd417767d5e5802c4"} Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.219641 4733 scope.go:117] "RemoveContainer" containerID="59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.219770 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmw2d" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.222606 4733 generic.go:334] "Generic (PLEG): container finished" podID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerID="e4094d8b4eb850df07d3a19e616d2e9c130ee64b58c626295f963df49e875ea5" exitCode=0 Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.222663 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerDied","Data":"e4094d8b4eb850df07d3a19e616d2e9c130ee64b58c626295f963df49e875ea5"} Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.230393 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx" (OuterVolumeSpecName: "kube-api-access-rs5rx") pod "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" (UID: "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc"). InnerVolumeSpecName "kube-api-access-rs5rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.231709 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerStarted","Data":"7e9cf80fc09f50439f722c47d01b38f7d154cd5514d553f0573a5303858564f5"} Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.264814 4733 scope.go:117] "RemoveContainer" containerID="c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.280558 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rls2r" podStartSLOduration=4.117886001 podStartE2EDuration="57.280534814s" podCreationTimestamp="2026-03-18 10:16:33 +0000 UTC" firstStartedPulling="2026-03-18 10:16:35.749727766 +0000 UTC m=+235.241462091" lastFinishedPulling="2026-03-18 10:17:28.912376579 +0000 UTC m=+288.404110904" observedRunningTime="2026-03-18 10:17:30.272480679 +0000 UTC m=+289.764215014" watchObservedRunningTime="2026-03-18 10:17:30.280534814 +0000 UTC m=+289.772269139" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.285727 4733 scope.go:117] "RemoveContainer" containerID="17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.286176 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" (UID: "7eb97f2d-18fa-4e8c-895f-de4602c9dbbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.301039 4733 scope.go:117] "RemoveContainer" containerID="59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d" Mar 18 10:17:30 crc kubenswrapper[4733]: E0318 10:17:30.301757 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d\": container with ID starting with 59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d not found: ID does not exist" containerID="59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.301823 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d"} err="failed to get container status \"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d\": rpc error: code = NotFound desc = could not find container \"59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d\": container with ID starting with 59fff67ee2ee5eba64f97e1503253ee04deab9883241245399247be534c88d2d not found: ID does not exist" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.301863 4733 scope.go:117] "RemoveContainer" containerID="c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c" Mar 18 10:17:30 crc kubenswrapper[4733]: E0318 10:17:30.302389 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c\": container with ID starting with c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c not found: ID does not exist" containerID="c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.302463 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c"} err="failed to get container status \"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c\": rpc error: code = NotFound desc = could not find container \"c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c\": container with ID starting with c0f64ebe0af1fce843609f30c833e4b965000df30afbd1af8fae99160a42210c not found: ID does not exist" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.302498 4733 scope.go:117] "RemoveContainer" containerID="17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266" Mar 18 10:17:30 crc kubenswrapper[4733]: E0318 10:17:30.302878 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266\": container with ID starting with 17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266 not found: ID does not exist" containerID="17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.302947 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266"} err="failed to get container status \"17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266\": rpc error: code = NotFound desc = could not find container \"17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266\": container with ID starting with 17741288ba852c25d8355eb97aa338d2e36690e9d066bbb56a0857710c52f266 not found: ID does not exist" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.317831 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.317856 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.317868 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs5rx\" (UniqueName: \"kubernetes.io/projected/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc-kube-api-access-rs5rx\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.554155 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:17:30 crc kubenswrapper[4733]: I0318 10:17:30.557766 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gmw2d"] Mar 18 10:17:31 crc kubenswrapper[4733]: I0318 10:17:31.184307 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" path="/var/lib/kubelet/pods/7eb97f2d-18fa-4e8c-895f-de4602c9dbbc/volumes" Mar 18 10:17:31 crc kubenswrapper[4733]: I0318 10:17:31.245379 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerStarted","Data":"ceccb7b8039b2e6bd282db7ca8b7756c23dd7eceb685ae3d99b11f442338e94b"} Mar 18 10:17:31 crc kubenswrapper[4733]: I0318 10:17:31.816930 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f82xf" podStartSLOduration=3.17356909 podStartE2EDuration="55.816904468s" podCreationTimestamp="2026-03-18 10:16:36 +0000 UTC" firstStartedPulling="2026-03-18 10:16:37.971698041 +0000 UTC m=+237.463432366" lastFinishedPulling="2026-03-18 10:17:30.615033419 +0000 UTC m=+290.106767744" observedRunningTime="2026-03-18 10:17:31.266377205 +0000 UTC m=+290.758111540" watchObservedRunningTime="2026-03-18 10:17:31.816904468 +0000 UTC m=+291.308638813" Mar 18 10:17:31 crc kubenswrapper[4733]: I0318 10:17:31.821796 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:17:31 crc kubenswrapper[4733]: I0318 10:17:31.822161 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ll7dp" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="registry-server" containerID="cri-o://7a59787487ae6ad4c8775bad7fe0e44e006e25d60fd069b4d5bd8cc6ceca6c70" gracePeriod=2 Mar 18 10:17:32 crc kubenswrapper[4733]: I0318 10:17:32.254430 4733 generic.go:334] "Generic (PLEG): container finished" podID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerID="7a59787487ae6ad4c8775bad7fe0e44e006e25d60fd069b4d5bd8cc6ceca6c70" exitCode=0 Mar 18 10:17:32 crc kubenswrapper[4733]: I0318 10:17:32.254517 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerDied","Data":"7a59787487ae6ad4c8775bad7fe0e44e006e25d60fd069b4d5bd8cc6ceca6c70"} Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.023871 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.129639 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.129944 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" podUID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" containerName="controller-manager" containerID="cri-o://c282a4dca9ca3d5260c23c43faf22ce611247dc38c8752227edd1fbf81bf90dd" gracePeriod=30 Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.148747 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.149052 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" podUID="debb4f43-7f9a-4fdc-9896-db5106650a74" containerName="route-controller-manager" containerID="cri-o://7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740" gracePeriod=30 Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.166775 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txwcn\" (UniqueName: \"kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn\") pod \"82922e1e-56fb-432e-9441-b99bdb19fbb0\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.168152 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities\") pod \"82922e1e-56fb-432e-9441-b99bdb19fbb0\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.168226 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content\") pod \"82922e1e-56fb-432e-9441-b99bdb19fbb0\" (UID: \"82922e1e-56fb-432e-9441-b99bdb19fbb0\") " Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.169102 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities" (OuterVolumeSpecName: "utilities") pod "82922e1e-56fb-432e-9441-b99bdb19fbb0" (UID: "82922e1e-56fb-432e-9441-b99bdb19fbb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.183363 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn" (OuterVolumeSpecName: "kube-api-access-txwcn") pod "82922e1e-56fb-432e-9441-b99bdb19fbb0" (UID: "82922e1e-56fb-432e-9441-b99bdb19fbb0"). InnerVolumeSpecName "kube-api-access-txwcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.262801 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll7dp" event={"ID":"82922e1e-56fb-432e-9441-b99bdb19fbb0","Type":"ContainerDied","Data":"1a1211028e93b8b114b76fa499d9200418412506c6795f17a8a464f56e421c4c"} Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.262866 4733 scope.go:117] "RemoveContainer" containerID="7a59787487ae6ad4c8775bad7fe0e44e006e25d60fd069b4d5bd8cc6ceca6c70" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.262989 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll7dp" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.269664 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.269686 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txwcn\" (UniqueName: \"kubernetes.io/projected/82922e1e-56fb-432e-9441-b99bdb19fbb0-kube-api-access-txwcn\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.282181 4733 scope.go:117] "RemoveContainer" containerID="29a9561b8927709b5dd59a92cbf81b78eacad45dd6ac5ec49191d6faee246d53" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.298274 4733 scope.go:117] "RemoveContainer" containerID="9d9d502e889f0bc1ff5ac5bd25eb5937fb15878b89bb5f2186b3e420cda96e62" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.309202 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82922e1e-56fb-432e-9441-b99bdb19fbb0" (UID: "82922e1e-56fb-432e-9441-b99bdb19fbb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.370767 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82922e1e-56fb-432e-9441-b99bdb19fbb0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.594064 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:17:33 crc kubenswrapper[4733]: I0318 10:17:33.605441 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ll7dp"] Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.143696 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.268910 4733 generic.go:334] "Generic (PLEG): container finished" podID="debb4f43-7f9a-4fdc-9896-db5106650a74" containerID="7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740" exitCode=0 Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.268999 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" event={"ID":"debb4f43-7f9a-4fdc-9896-db5106650a74","Type":"ContainerDied","Data":"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740"} Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.269035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" event={"ID":"debb4f43-7f9a-4fdc-9896-db5106650a74","Type":"ContainerDied","Data":"f30c8eb187bc3102906d7afcd8a11a3b91672d508cc51863174f4d3b52b10205"} Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.269056 4733 scope.go:117] "RemoveContainer" containerID="7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.269092 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.273353 4733 generic.go:334] "Generic (PLEG): container finished" podID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" containerID="c282a4dca9ca3d5260c23c43faf22ce611247dc38c8752227edd1fbf81bf90dd" exitCode=0 Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.273390 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" event={"ID":"f280b8dd-9b2f-4faa-9335-ca84961ea0a6","Type":"ContainerDied","Data":"c282a4dca9ca3d5260c23c43faf22ce611247dc38c8752227edd1fbf81bf90dd"} Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.282722 4733 scope.go:117] "RemoveContainer" containerID="7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.283184 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740\": container with ID starting with 7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740 not found: ID does not exist" containerID="7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.283290 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740"} err="failed to get container status \"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740\": rpc error: code = NotFound desc = could not find container \"7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740\": container with ID starting with 7412c3c1effa7cc3545719a633ad64080beb347c499c8ddc9951beee83c2e740 not found: ID does not exist" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.284716 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwfk4\" (UniqueName: \"kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4\") pod \"debb4f43-7f9a-4fdc-9896-db5106650a74\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.284803 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca\") pod \"debb4f43-7f9a-4fdc-9896-db5106650a74\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.284855 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config\") pod \"debb4f43-7f9a-4fdc-9896-db5106650a74\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.284944 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert\") pod \"debb4f43-7f9a-4fdc-9896-db5106650a74\" (UID: \"debb4f43-7f9a-4fdc-9896-db5106650a74\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.285938 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config" (OuterVolumeSpecName: "config") pod "debb4f43-7f9a-4fdc-9896-db5106650a74" (UID: "debb4f43-7f9a-4fdc-9896-db5106650a74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.286569 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca" (OuterVolumeSpecName: "client-ca") pod "debb4f43-7f9a-4fdc-9896-db5106650a74" (UID: "debb4f43-7f9a-4fdc-9896-db5106650a74"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.290148 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4" (OuterVolumeSpecName: "kube-api-access-wwfk4") pod "debb4f43-7f9a-4fdc-9896-db5106650a74" (UID: "debb4f43-7f9a-4fdc-9896-db5106650a74"). InnerVolumeSpecName "kube-api-access-wwfk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.290637 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "debb4f43-7f9a-4fdc-9896-db5106650a74" (UID: "debb4f43-7f9a-4fdc-9896-db5106650a74"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.324818 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.386807 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config\") pod \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.386865 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-575kc\" (UniqueName: \"kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc\") pod \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.386919 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca\") pod \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387016 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles\") pod \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387045 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert\") pod \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\" (UID: \"f280b8dd-9b2f-4faa-9335-ca84961ea0a6\") " Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387304 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwfk4\" (UniqueName: \"kubernetes.io/projected/debb4f43-7f9a-4fdc-9896-db5106650a74-kube-api-access-wwfk4\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387319 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387329 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/debb4f43-7f9a-4fdc-9896-db5106650a74-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.387337 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/debb4f43-7f9a-4fdc-9896-db5106650a74-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.388113 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "f280b8dd-9b2f-4faa-9335-ca84961ea0a6" (UID: "f280b8dd-9b2f-4faa-9335-ca84961ea0a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.388231 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f280b8dd-9b2f-4faa-9335-ca84961ea0a6" (UID: "f280b8dd-9b2f-4faa-9335-ca84961ea0a6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.388842 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config" (OuterVolumeSpecName: "config") pod "f280b8dd-9b2f-4faa-9335-ca84961ea0a6" (UID: "f280b8dd-9b2f-4faa-9335-ca84961ea0a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.390419 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f280b8dd-9b2f-4faa-9335-ca84961ea0a6" (UID: "f280b8dd-9b2f-4faa-9335-ca84961ea0a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.390476 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc" (OuterVolumeSpecName: "kube-api-access-575kc") pod "f280b8dd-9b2f-4faa-9335-ca84961ea0a6" (UID: "f280b8dd-9b2f-4faa-9335-ca84961ea0a6"). InnerVolumeSpecName "kube-api-access-575kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.429122 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.429394 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.451067 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452085 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" containerName="controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452097 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" containerName="controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452108 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="extract-utilities" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452114 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="extract-utilities" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452121 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="extract-utilities" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452127 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="extract-utilities" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452137 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452142 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452149 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="extract-content" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452155 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="extract-content" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452164 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="extract-content" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452169 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="extract-content" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452176 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452181 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: E0318 10:17:34.452201 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="debb4f43-7f9a-4fdc-9896-db5106650a74" containerName="route-controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452206 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="debb4f43-7f9a-4fdc-9896-db5106650a74" containerName="route-controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452303 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb97f2d-18fa-4e8c-895f-de4602c9dbbc" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452314 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" containerName="registry-server" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452322 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="debb4f43-7f9a-4fdc-9896-db5106650a74" containerName="route-controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452331 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" containerName="controller-manager" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.452660 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.467257 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.476092 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.488432 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.488464 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-575kc\" (UniqueName: \"kubernetes.io/projected/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-kube-api-access-575kc\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.488474 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.488483 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.488491 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f280b8dd-9b2f-4faa-9335-ca84961ea0a6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.589378 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgfsn\" (UniqueName: \"kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.589533 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.589555 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.589603 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.593344 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.596840 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6966b4866d-xm6hn"] Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.690605 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.690654 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.690695 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.690762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgfsn\" (UniqueName: \"kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.691760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.692075 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.698858 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.705857 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgfsn\" (UniqueName: \"kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn\") pod \"route-controller-manager-795c5666f8-fqxxn\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.723499 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.723529 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.763519 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.767764 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:34 crc kubenswrapper[4733]: I0318 10:17:34.962917 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:34 crc kubenswrapper[4733]: W0318 10:17:34.971061 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2338c705_9627_4c7c_97c5_60c492309e8f.slice/crio-d2316c222486d15d81f6544c130d255f209d75ef4a225b569454d5bafde3bc6e WatchSource:0}: Error finding container d2316c222486d15d81f6544c130d255f209d75ef4a225b569454d5bafde3bc6e: Status 404 returned error can't find the container with id d2316c222486d15d81f6544c130d255f209d75ef4a225b569454d5bafde3bc6e Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.183645 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82922e1e-56fb-432e-9441-b99bdb19fbb0" path="/var/lib/kubelet/pods/82922e1e-56fb-432e-9441-b99bdb19fbb0/volumes" Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.185356 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="debb4f43-7f9a-4fdc-9896-db5106650a74" path="/var/lib/kubelet/pods/debb4f43-7f9a-4fdc-9896-db5106650a74/volumes" Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.281046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" event={"ID":"f280b8dd-9b2f-4faa-9335-ca84961ea0a6","Type":"ContainerDied","Data":"15b644538c3645c79d06f703952a07b7b4ae39d43d1dc32a31009b361d9095cd"} Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.281061 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cf89766c-wb795" Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.281115 4733 scope.go:117] "RemoveContainer" containerID="c282a4dca9ca3d5260c23c43faf22ce611247dc38c8752227edd1fbf81bf90dd" Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.282507 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" event={"ID":"2338c705-9627-4c7c-97c5-60c492309e8f","Type":"ContainerStarted","Data":"d2316c222486d15d81f6544c130d255f209d75ef4a225b569454d5bafde3bc6e"} Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.306258 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.308711 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cf89766c-wb795"] Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.320807 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:35 crc kubenswrapper[4733]: I0318 10:17:35.331402 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.243384 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.243517 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.289035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.290821 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" event={"ID":"2338c705-9627-4c7c-97c5-60c492309e8f","Type":"ContainerStarted","Data":"b5cbc2cf0ca3b801430e89c3bd7280dd3b0874c8d8c15426c4e2737151b8f3ad"} Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.291079 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.296788 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.340559 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.433604 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" podStartSLOduration=3.433581354 podStartE2EDuration="3.433581354s" podCreationTimestamp="2026-03-18 10:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:36.349178901 +0000 UTC m=+295.840913216" watchObservedRunningTime="2026-03-18 10:17:36.433581354 +0000 UTC m=+295.925315679" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.454847 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.456275 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.462795 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.464229 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.464443 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.464541 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.464227 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.464813 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.471349 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.480831 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.521650 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.521711 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.521737 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bffk7\" (UniqueName: \"kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.521762 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.521956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.623603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.623664 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.623686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bffk7\" (UniqueName: \"kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.623713 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.623759 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.624952 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.625068 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.625164 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.631613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.633089 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.633128 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.648935 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bffk7\" (UniqueName: \"kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7\") pod \"controller-manager-855cb9cb7d-nd8zj\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.688140 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.774601 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:36 crc kubenswrapper[4733]: I0318 10:17:36.981970 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.183283 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f280b8dd-9b2f-4faa-9335-ca84961ea0a6" path="/var/lib/kubelet/pods/f280b8dd-9b2f-4faa-9335-ca84961ea0a6/volumes" Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.300277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" event={"ID":"9c910a75-c5cb-4f2e-ba5b-29866e412aae","Type":"ContainerStarted","Data":"b2763b1f739435470aef969e3ffa833e68884b68a2e5a4fe94dd1c73a705f26a"} Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.300341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" event={"ID":"9c910a75-c5cb-4f2e-ba5b-29866e412aae","Type":"ContainerStarted","Data":"429b52d24bd3668b5415677a4bed9fb2a028c9049e40bf9e468c59dfe2fee20b"} Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.347835 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.381277 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" podStartSLOduration=4.381237756 podStartE2EDuration="4.381237756s" podCreationTimestamp="2026-03-18 10:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:37.328972333 +0000 UTC m=+296.820706658" watchObservedRunningTime="2026-03-18 10:17:37.381237756 +0000 UTC m=+296.872972071" Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.619370 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:17:37 crc kubenswrapper[4733]: I0318 10:17:37.619626 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w7rrs" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="registry-server" containerID="cri-o://0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6" gracePeriod=2 Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.035973 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.146043 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities\") pod \"02cd6358-355c-4db8-b0f7-2528618602ff\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.146167 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5vpf\" (UniqueName: \"kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf\") pod \"02cd6358-355c-4db8-b0f7-2528618602ff\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.146288 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content\") pod \"02cd6358-355c-4db8-b0f7-2528618602ff\" (UID: \"02cd6358-355c-4db8-b0f7-2528618602ff\") " Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.147465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities" (OuterVolumeSpecName: "utilities") pod "02cd6358-355c-4db8-b0f7-2528618602ff" (UID: "02cd6358-355c-4db8-b0f7-2528618602ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.156182 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf" (OuterVolumeSpecName: "kube-api-access-s5vpf") pod "02cd6358-355c-4db8-b0f7-2528618602ff" (UID: "02cd6358-355c-4db8-b0f7-2528618602ff"). InnerVolumeSpecName "kube-api-access-s5vpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.193372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02cd6358-355c-4db8-b0f7-2528618602ff" (UID: "02cd6358-355c-4db8-b0f7-2528618602ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.248485 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.248539 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cd6358-355c-4db8-b0f7-2528618602ff-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.248565 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5vpf\" (UniqueName: \"kubernetes.io/projected/02cd6358-355c-4db8-b0f7-2528618602ff-kube-api-access-s5vpf\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.306713 4733 generic.go:334] "Generic (PLEG): container finished" podID="02cd6358-355c-4db8-b0f7-2528618602ff" containerID="0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6" exitCode=0 Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.307033 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7rrs" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.307078 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerDied","Data":"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6"} Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.307162 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7rrs" event={"ID":"02cd6358-355c-4db8-b0f7-2528618602ff","Type":"ContainerDied","Data":"bf716e26a7a1e4408c9cf17e7366833bdc30d38efd823adf2eb5d92d8a80e381"} Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.307201 4733 scope.go:117] "RemoveContainer" containerID="0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.308015 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.314605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.337368 4733 scope.go:117] "RemoveContainer" containerID="6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.360856 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.360916 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w7rrs"] Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.374425 4733 scope.go:117] "RemoveContainer" containerID="0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.400342 4733 scope.go:117] "RemoveContainer" containerID="0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6" Mar 18 10:17:38 crc kubenswrapper[4733]: E0318 10:17:38.400771 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6\": container with ID starting with 0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6 not found: ID does not exist" containerID="0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.400809 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6"} err="failed to get container status \"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6\": rpc error: code = NotFound desc = could not find container \"0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6\": container with ID starting with 0670cad9cd6cfa6a4cf42522d884617aa8495c18fcee7ac17083576de64388b6 not found: ID does not exist" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.400836 4733 scope.go:117] "RemoveContainer" containerID="6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450" Mar 18 10:17:38 crc kubenswrapper[4733]: E0318 10:17:38.401555 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450\": container with ID starting with 6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450 not found: ID does not exist" containerID="6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.401576 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450"} err="failed to get container status \"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450\": rpc error: code = NotFound desc = could not find container \"6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450\": container with ID starting with 6888d6a2cdf0ef31a9d456bbbbc7efb04bffa6fcf33a7a14044f9c00de4a1450 not found: ID does not exist" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.401591 4733 scope.go:117] "RemoveContainer" containerID="0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed" Mar 18 10:17:38 crc kubenswrapper[4733]: E0318 10:17:38.402061 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed\": container with ID starting with 0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed not found: ID does not exist" containerID="0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed" Mar 18 10:17:38 crc kubenswrapper[4733]: I0318 10:17:38.402119 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed"} err="failed to get container status \"0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed\": rpc error: code = NotFound desc = could not find container \"0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed\": container with ID starting with 0f769f3a01023165d2b55b37631a8e64c99ba9561927f3f83986829531dcb6ed not found: ID does not exist" Mar 18 10:17:39 crc kubenswrapper[4733]: I0318 10:17:39.182326 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" path="/var/lib/kubelet/pods/02cd6358-355c-4db8-b0f7-2528618602ff/volumes" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.025020 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.025464 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f82xf" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="registry-server" containerID="cri-o://ceccb7b8039b2e6bd282db7ca8b7756c23dd7eceb685ae3d99b11f442338e94b" gracePeriod=2 Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.324009 4733 generic.go:334] "Generic (PLEG): container finished" podID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerID="ceccb7b8039b2e6bd282db7ca8b7756c23dd7eceb685ae3d99b11f442338e94b" exitCode=0 Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.324084 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerDied","Data":"ceccb7b8039b2e6bd282db7ca8b7756c23dd7eceb685ae3d99b11f442338e94b"} Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.541018 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.733450 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content\") pod \"c91f12fa-96f0-442a-a3f7-70d56a697839\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.733529 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities\") pod \"c91f12fa-96f0-442a-a3f7-70d56a697839\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.733610 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r2vl\" (UniqueName: \"kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl\") pod \"c91f12fa-96f0-442a-a3f7-70d56a697839\" (UID: \"c91f12fa-96f0-442a-a3f7-70d56a697839\") " Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.734755 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities" (OuterVolumeSpecName: "utilities") pod "c91f12fa-96f0-442a-a3f7-70d56a697839" (UID: "c91f12fa-96f0-442a-a3f7-70d56a697839"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.741046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl" (OuterVolumeSpecName: "kube-api-access-6r2vl") pod "c91f12fa-96f0-442a-a3f7-70d56a697839" (UID: "c91f12fa-96f0-442a-a3f7-70d56a697839"). InnerVolumeSpecName "kube-api-access-6r2vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.831627 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c91f12fa-96f0-442a-a3f7-70d56a697839" (UID: "c91f12fa-96f0-442a-a3f7-70d56a697839"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.834832 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r2vl\" (UniqueName: \"kubernetes.io/projected/c91f12fa-96f0-442a-a3f7-70d56a697839-kube-api-access-6r2vl\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.834867 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:40 crc kubenswrapper[4733]: I0318 10:17:40.834879 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c91f12fa-96f0-442a-a3f7-70d56a697839-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.334095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f82xf" event={"ID":"c91f12fa-96f0-442a-a3f7-70d56a697839","Type":"ContainerDied","Data":"a8fa061a3aa824aa80f6c1569abe326d18dccd731789c62f81d22de7e9a828d3"} Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.334291 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f82xf" Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.334583 4733 scope.go:117] "RemoveContainer" containerID="ceccb7b8039b2e6bd282db7ca8b7756c23dd7eceb685ae3d99b11f442338e94b" Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.365381 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.371874 4733 scope.go:117] "RemoveContainer" containerID="e4094d8b4eb850df07d3a19e616d2e9c130ee64b58c626295f963df49e875ea5" Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.372697 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f82xf"] Mar 18 10:17:41 crc kubenswrapper[4733]: I0318 10:17:41.398234 4733 scope.go:117] "RemoveContainer" containerID="6da3522bbcdb557467c36bac266a9dafb390a5a917de44dd30de9c3ac03051e1" Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.186429 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" path="/var/lib/kubelet/pods/c91f12fa-96f0-442a-a3f7-70d56a697839/volumes" Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.571123 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.572089 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.572164 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.572923 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:17:43 crc kubenswrapper[4733]: I0318 10:17:43.572987 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830" gracePeriod=600 Mar 18 10:17:44 crc kubenswrapper[4733]: I0318 10:17:44.361637 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830" exitCode=0 Mar 18 10:17:44 crc kubenswrapper[4733]: I0318 10:17:44.362043 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830"} Mar 18 10:17:44 crc kubenswrapper[4733]: I0318 10:17:44.362090 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028"} Mar 18 10:17:46 crc kubenswrapper[4733]: I0318 10:17:46.385848 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-n6hmz"] Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.109332 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.110199 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" podUID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" containerName="controller-manager" containerID="cri-o://b2763b1f739435470aef969e3ffa833e68884b68a2e5a4fe94dd1c73a705f26a" gracePeriod=30 Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.201721 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.201983 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" podUID="2338c705-9627-4c7c-97c5-60c492309e8f" containerName="route-controller-manager" containerID="cri-o://b5cbc2cf0ca3b801430e89c3bd7280dd3b0874c8d8c15426c4e2737151b8f3ad" gracePeriod=30 Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.419406 4733 generic.go:334] "Generic (PLEG): container finished" podID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" containerID="b2763b1f739435470aef969e3ffa833e68884b68a2e5a4fe94dd1c73a705f26a" exitCode=0 Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.419474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" event={"ID":"9c910a75-c5cb-4f2e-ba5b-29866e412aae","Type":"ContainerDied","Data":"b2763b1f739435470aef969e3ffa833e68884b68a2e5a4fe94dd1c73a705f26a"} Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.421715 4733 generic.go:334] "Generic (PLEG): container finished" podID="2338c705-9627-4c7c-97c5-60c492309e8f" containerID="b5cbc2cf0ca3b801430e89c3bd7280dd3b0874c8d8c15426c4e2737151b8f3ad" exitCode=0 Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.421761 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" event={"ID":"2338c705-9627-4c7c-97c5-60c492309e8f","Type":"ContainerDied","Data":"b5cbc2cf0ca3b801430e89c3bd7280dd3b0874c8d8c15426c4e2737151b8f3ad"} Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.687568 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.695440 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826099 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert\") pod \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config\") pod \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826227 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config\") pod \"2338c705-9627-4c7c-97c5-60c492309e8f\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826287 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca\") pod \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826311 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert\") pod \"2338c705-9627-4c7c-97c5-60c492309e8f\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826339 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bffk7\" (UniqueName: \"kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7\") pod \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles\") pod \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\" (UID: \"9c910a75-c5cb-4f2e-ba5b-29866e412aae\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826453 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca\") pod \"2338c705-9627-4c7c-97c5-60c492309e8f\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.826485 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgfsn\" (UniqueName: \"kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn\") pod \"2338c705-9627-4c7c-97c5-60c492309e8f\" (UID: \"2338c705-9627-4c7c-97c5-60c492309e8f\") " Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.827353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9c910a75-c5cb-4f2e-ba5b-29866e412aae" (UID: "9c910a75-c5cb-4f2e-ba5b-29866e412aae"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.827374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca" (OuterVolumeSpecName: "client-ca") pod "9c910a75-c5cb-4f2e-ba5b-29866e412aae" (UID: "9c910a75-c5cb-4f2e-ba5b-29866e412aae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.827353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca" (OuterVolumeSpecName: "client-ca") pod "2338c705-9627-4c7c-97c5-60c492309e8f" (UID: "2338c705-9627-4c7c-97c5-60c492309e8f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.827598 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config" (OuterVolumeSpecName: "config") pod "2338c705-9627-4c7c-97c5-60c492309e8f" (UID: "2338c705-9627-4c7c-97c5-60c492309e8f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.827802 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config" (OuterVolumeSpecName: "config") pod "9c910a75-c5cb-4f2e-ba5b-29866e412aae" (UID: "9c910a75-c5cb-4f2e-ba5b-29866e412aae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.832273 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7" (OuterVolumeSpecName: "kube-api-access-bffk7") pod "9c910a75-c5cb-4f2e-ba5b-29866e412aae" (UID: "9c910a75-c5cb-4f2e-ba5b-29866e412aae"). InnerVolumeSpecName "kube-api-access-bffk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.832453 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9c910a75-c5cb-4f2e-ba5b-29866e412aae" (UID: "9c910a75-c5cb-4f2e-ba5b-29866e412aae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.832494 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn" (OuterVolumeSpecName: "kube-api-access-sgfsn") pod "2338c705-9627-4c7c-97c5-60c492309e8f" (UID: "2338c705-9627-4c7c-97c5-60c492309e8f"). InnerVolumeSpecName "kube-api-access-sgfsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.832616 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2338c705-9627-4c7c-97c5-60c492309e8f" (UID: "2338c705-9627-4c7c-97c5-60c492309e8f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.927882 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c910a75-c5cb-4f2e-ba5b-29866e412aae-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928090 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928213 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928287 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928348 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2338c705-9627-4c7c-97c5-60c492309e8f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928409 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bffk7\" (UniqueName: \"kubernetes.io/projected/9c910a75-c5cb-4f2e-ba5b-29866e412aae-kube-api-access-bffk7\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928632 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c910a75-c5cb-4f2e-ba5b-29866e412aae-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928674 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2338c705-9627-4c7c-97c5-60c492309e8f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:53 crc kubenswrapper[4733]: I0318 10:17:53.928693 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgfsn\" (UniqueName: \"kubernetes.io/projected/2338c705-9627-4c7c-97c5-60c492309e8f-kube-api-access-sgfsn\") on node \"crc\" DevicePath \"\"" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.431540 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.431562 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn" event={"ID":"2338c705-9627-4c7c-97c5-60c492309e8f","Type":"ContainerDied","Data":"d2316c222486d15d81f6544c130d255f209d75ef4a225b569454d5bafde3bc6e"} Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.431670 4733 scope.go:117] "RemoveContainer" containerID="b5cbc2cf0ca3b801430e89c3bd7280dd3b0874c8d8c15426c4e2737151b8f3ad" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.438009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" event={"ID":"9c910a75-c5cb-4f2e-ba5b-29866e412aae","Type":"ContainerDied","Data":"429b52d24bd3668b5415677a4bed9fb2a028c9049e40bf9e468c59dfe2fee20b"} Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.438167 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.451401 4733 scope.go:117] "RemoveContainer" containerID="b2763b1f739435470aef969e3ffa833e68884b68a2e5a4fe94dd1c73a705f26a" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.467039 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56547f49c8-qcl2v"] Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.467992 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="extract-utilities" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.468254 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="extract-utilities" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.468440 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.468678 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.468873 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.469050 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.469276 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="extract-utilities" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.469458 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="extract-utilities" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.469666 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="extract-content" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.472552 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="extract-content" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.472849 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2338c705-9627-4c7c-97c5-60c492309e8f" containerName="route-controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.473019 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2338c705-9627-4c7c-97c5-60c492309e8f" containerName="route-controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.473164 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" containerName="controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.473743 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" containerName="controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: E0318 10:17:54.473923 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="extract-content" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.474096 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="extract-content" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.474786 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" containerName="controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.474837 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="02cd6358-355c-4db8-b0f7-2528618602ff" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.474850 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91f12fa-96f0-442a-a3f7-70d56a697839" containerName="registry-server" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.474871 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2338c705-9627-4c7c-97c5-60c492309e8f" containerName="route-controller-manager" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.476442 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.477237 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.478038 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.481755 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.482763 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.485832 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.487289 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.487622 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488325 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488556 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488678 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488744 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488900 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488950 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.488550 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.489257 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.490294 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795c5666f8-fqxxn"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.494603 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.495650 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56547f49c8-qcl2v"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.498222 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.536604 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-config\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.536664 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-proxy-ca-bundles\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.536806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv8sx\" (UniqueName: \"kubernetes.io/projected/f128aaea-3d79-459f-9a33-3a2505089c1c-kube-api-access-cv8sx\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.536881 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-client-ca\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.536965 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-config\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.537087 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjfg9\" (UniqueName: \"kubernetes.io/projected/6ff40242-1a33-443f-892d-145e88f98be2-kube-api-access-qjfg9\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.537115 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-client-ca\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.537138 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ff40242-1a33-443f-892d-145e88f98be2-serving-cert\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.537169 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f128aaea-3d79-459f-9a33-3a2505089c1c-serving-cert\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.539783 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.548364 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-855cb9cb7d-nd8zj"] Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638448 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-client-ca\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638548 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ff40242-1a33-443f-892d-145e88f98be2-serving-cert\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638735 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f128aaea-3d79-459f-9a33-3a2505089c1c-serving-cert\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638816 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-config\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-proxy-ca-bundles\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv8sx\" (UniqueName: \"kubernetes.io/projected/f128aaea-3d79-459f-9a33-3a2505089c1c-kube-api-access-cv8sx\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.638993 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-client-ca\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.639096 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-config\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.639322 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjfg9\" (UniqueName: \"kubernetes.io/projected/6ff40242-1a33-443f-892d-145e88f98be2-kube-api-access-qjfg9\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.639945 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-client-ca\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.641236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-client-ca\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.641340 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-proxy-ca-bundles\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.642137 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ff40242-1a33-443f-892d-145e88f98be2-config\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.642466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f128aaea-3d79-459f-9a33-3a2505089c1c-config\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.644580 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f128aaea-3d79-459f-9a33-3a2505089c1c-serving-cert\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.648738 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ff40242-1a33-443f-892d-145e88f98be2-serving-cert\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.659736 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjfg9\" (UniqueName: \"kubernetes.io/projected/6ff40242-1a33-443f-892d-145e88f98be2-kube-api-access-qjfg9\") pod \"controller-manager-56547f49c8-qcl2v\" (UID: \"6ff40242-1a33-443f-892d-145e88f98be2\") " pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.661960 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv8sx\" (UniqueName: \"kubernetes.io/projected/f128aaea-3d79-459f-9a33-3a2505089c1c-kube-api-access-cv8sx\") pod \"route-controller-manager-5975b786b-2xcz5\" (UID: \"f128aaea-3d79-459f-9a33-3a2505089c1c\") " pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.845941 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:54 crc kubenswrapper[4733]: I0318 10:17:54.856171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.090410 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56547f49c8-qcl2v"] Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.184518 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2338c705-9627-4c7c-97c5-60c492309e8f" path="/var/lib/kubelet/pods/2338c705-9627-4c7c-97c5-60c492309e8f/volumes" Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.185400 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c910a75-c5cb-4f2e-ba5b-29866e412aae" path="/var/lib/kubelet/pods/9c910a75-c5cb-4f2e-ba5b-29866e412aae/volumes" Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.254357 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5"] Mar 18 10:17:55 crc kubenswrapper[4733]: W0318 10:17:55.263615 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf128aaea_3d79_459f_9a33_3a2505089c1c.slice/crio-cdf7013e2f2b84291c4f32f283296d753d937c2f51fa8a1530466e4ba3550edd WatchSource:0}: Error finding container cdf7013e2f2b84291c4f32f283296d753d937c2f51fa8a1530466e4ba3550edd: Status 404 returned error can't find the container with id cdf7013e2f2b84291c4f32f283296d753d937c2f51fa8a1530466e4ba3550edd Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.446112 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" event={"ID":"f128aaea-3d79-459f-9a33-3a2505089c1c","Type":"ContainerStarted","Data":"cdf7013e2f2b84291c4f32f283296d753d937c2f51fa8a1530466e4ba3550edd"} Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.449406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" event={"ID":"6ff40242-1a33-443f-892d-145e88f98be2","Type":"ContainerStarted","Data":"720c9ebb3f4e168d5a9de625b15345ed0376c77651ce2ea95b52e8d23eeeb0ca"} Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.449458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" event={"ID":"6ff40242-1a33-443f-892d-145e88f98be2","Type":"ContainerStarted","Data":"df50f7e523e0addea33cd24a662f83fd05891abc2d4d7482d097e979154a9807"} Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.450571 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.454223 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" Mar 18 10:17:55 crc kubenswrapper[4733]: I0318 10:17:55.466819 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56547f49c8-qcl2v" podStartSLOduration=2.466805775 podStartE2EDuration="2.466805775s" podCreationTimestamp="2026-03-18 10:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:55.465957992 +0000 UTC m=+314.957692317" watchObservedRunningTime="2026-03-18 10:17:55.466805775 +0000 UTC m=+314.958540090" Mar 18 10:17:56 crc kubenswrapper[4733]: I0318 10:17:56.460174 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" event={"ID":"f128aaea-3d79-459f-9a33-3a2505089c1c","Type":"ContainerStarted","Data":"fc309deba97733a5605b5f9ba276b7e31a6f5046e89882a706264a8983d5cffe"} Mar 18 10:17:56 crc kubenswrapper[4733]: I0318 10:17:56.477733 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" podStartSLOduration=3.477714958 podStartE2EDuration="3.477714958s" podCreationTimestamp="2026-03-18 10:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:17:56.474731634 +0000 UTC m=+315.966465969" watchObservedRunningTime="2026-03-18 10:17:56.477714958 +0000 UTC m=+315.969449283" Mar 18 10:17:57 crc kubenswrapper[4733]: I0318 10:17:57.464685 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:57 crc kubenswrapper[4733]: I0318 10:17:57.469978 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5975b786b-2xcz5" Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.975306 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.982343 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.982991 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.983032 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba" gracePeriod=15 Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.983667 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052" gracePeriod=15 Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.983878 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6" gracePeriod=15 Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.983957 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa" gracePeriod=15 Mar 18 10:17:59 crc kubenswrapper[4733]: I0318 10:17:59.984550 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0" gracePeriod=15 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.003131 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.003918 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.003953 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.003979 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.003993 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004009 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004022 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004037 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004050 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004093 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004105 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004121 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004133 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004156 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004168 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004220 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004238 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.004268 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004281 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004573 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004606 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004629 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004656 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004672 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004689 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004714 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.004779 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 10:18:00 crc kubenswrapper[4733]: E0318 10:18:00.005152 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.005170 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.010412 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119323 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119363 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119394 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119465 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119508 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.119603 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220211 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220381 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220399 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220416 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.220507 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221161 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221453 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221490 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221699 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221723 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.221784 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.485236 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.488124 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.490659 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052" exitCode=0 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.490731 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6" exitCode=0 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.490750 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0" exitCode=0 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.490767 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa" exitCode=2 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.490879 4733 scope.go:117] "RemoveContainer" containerID="ba371d0dc81f8827d305037cab25306e3abe8ed3d243f74923b4709198f7ea38" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.494265 4733 generic.go:334] "Generic (PLEG): container finished" podID="71bc6618-8df4-4a35-9469-772a853eff06" containerID="7f5d2a4800b0b935a593d622bab229709f8902d75a6b9d3d310047bf50063a1a" exitCode=0 Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.494351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"71bc6618-8df4-4a35-9469-772a853eff06","Type":"ContainerDied","Data":"7f5d2a4800b0b935a593d622bab229709f8902d75a6b9d3d310047bf50063a1a"} Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.495599 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:00 crc kubenswrapper[4733]: I0318 10:18:00.496102 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.179108 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.179540 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.507209 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.923438 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.924170 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946264 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir\") pod \"71bc6618-8df4-4a35-9469-772a853eff06\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946417 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock\") pod \"71bc6618-8df4-4a35-9469-772a853eff06\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946416 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "71bc6618-8df4-4a35-9469-772a853eff06" (UID: "71bc6618-8df4-4a35-9469-772a853eff06"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946446 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access\") pod \"71bc6618-8df4-4a35-9469-772a853eff06\" (UID: \"71bc6618-8df4-4a35-9469-772a853eff06\") " Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946468 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock" (OuterVolumeSpecName: "var-lock") pod "71bc6618-8df4-4a35-9469-772a853eff06" (UID: "71bc6618-8df4-4a35-9469-772a853eff06"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946766 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.946777 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/71bc6618-8df4-4a35-9469-772a853eff06-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:01 crc kubenswrapper[4733]: I0318 10:18:01.952958 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "71bc6618-8df4-4a35-9469-772a853eff06" (UID: "71bc6618-8df4-4a35-9469-772a853eff06"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.048226 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71bc6618-8df4-4a35-9469-772a853eff06-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.370074 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.371069 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.371737 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.372218 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.451705 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.451766 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.451820 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.451877 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.451972 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.452169 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.452629 4733 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.452690 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.452711 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.517054 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.517943 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba" exitCode=0 Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.518010 4733 scope.go:117] "RemoveContainer" containerID="6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.518030 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.519845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"71bc6618-8df4-4a35-9469-772a853eff06","Type":"ContainerDied","Data":"5f44602afd79b72c25bccf945c72fd688dbc42ff6b86533bf0722398bd85fb3d"} Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.519888 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f44602afd79b72c25bccf945c72fd688dbc42ff6b86533bf0722398bd85fb3d" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.519975 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.532848 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.533573 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.539788 4733 scope.go:117] "RemoveContainer" containerID="b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.544842 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.545576 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.554308 4733 scope.go:117] "RemoveContainer" containerID="7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.574410 4733 scope.go:117] "RemoveContainer" containerID="edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.595901 4733 scope.go:117] "RemoveContainer" containerID="1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.617242 4733 scope.go:117] "RemoveContainer" containerID="ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.637636 4733 scope.go:117] "RemoveContainer" containerID="6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.638284 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\": container with ID starting with 6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052 not found: ID does not exist" containerID="6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.638361 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052"} err="failed to get container status \"6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\": rpc error: code = NotFound desc = could not find container \"6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052\": container with ID starting with 6fa9eed1a11fd6a14b82ea9f34ead9b9c67e9c9d52c2675651b37f9838875052 not found: ID does not exist" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.638413 4733 scope.go:117] "RemoveContainer" containerID="b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.638873 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\": container with ID starting with b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6 not found: ID does not exist" containerID="b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.638917 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6"} err="failed to get container status \"b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\": rpc error: code = NotFound desc = could not find container \"b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6\": container with ID starting with b698902beccdf67c5646c01b34eea131f61dee8d5d6e1f566cdb70c930b2cde6 not found: ID does not exist" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.638946 4733 scope.go:117] "RemoveContainer" containerID="7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.639386 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\": container with ID starting with 7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0 not found: ID does not exist" containerID="7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.639441 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0"} err="failed to get container status \"7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\": rpc error: code = NotFound desc = could not find container \"7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0\": container with ID starting with 7aaa002cf5203102149456e58fcc5db02a5e861736d3699e432a91186bac47d0 not found: ID does not exist" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.639509 4733 scope.go:117] "RemoveContainer" containerID="edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.639928 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\": container with ID starting with edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa not found: ID does not exist" containerID="edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.639989 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa"} err="failed to get container status \"edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\": rpc error: code = NotFound desc = could not find container \"edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa\": container with ID starting with edcafff0c9902e275fc23a2f154d3030c0e751e2f3230a4ca226c9cef8efcbfa not found: ID does not exist" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.640010 4733 scope.go:117] "RemoveContainer" containerID="1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.640319 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\": container with ID starting with 1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba not found: ID does not exist" containerID="1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.640346 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba"} err="failed to get container status \"1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\": rpc error: code = NotFound desc = could not find container \"1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba\": container with ID starting with 1614bd2915eb4ab62554cfe72d63669c062baaf25ae2e533788b876ff9544eba not found: ID does not exist" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.640364 4733 scope.go:117] "RemoveContainer" containerID="ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734" Mar 18 10:18:02 crc kubenswrapper[4733]: E0318 10:18:02.640671 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\": container with ID starting with ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734 not found: ID does not exist" containerID="ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734" Mar 18 10:18:02 crc kubenswrapper[4733]: I0318 10:18:02.640696 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734"} err="failed to get container status \"ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\": rpc error: code = NotFound desc = could not find container \"ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734\": container with ID starting with ea9fbcd7d532de4f7ea45ab0610692732e5b6f4df725aac2f68c72d8dbdb6734 not found: ID does not exist" Mar 18 10:18:03 crc kubenswrapper[4733]: I0318 10:18:03.190888 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.046157 4733 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:05 crc kubenswrapper[4733]: I0318 10:18:05.047129 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.083341 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189de8279c9a6215 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:18:05.082575381 +0000 UTC m=+324.574309706,LastTimestamp:2026-03-18 10:18:05.082575381 +0000 UTC m=+324.574309706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:18:05 crc kubenswrapper[4733]: I0318 10:18:05.543289 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0"} Mar 18 10:18:05 crc kubenswrapper[4733]: I0318 10:18:05.543355 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e08d1b29840f18c35701cc9dbeb6a5d031bf4e9b063c0a1576f871038a4ba2b5"} Mar 18 10:18:05 crc kubenswrapper[4733]: I0318 10:18:05.544061 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.544073 4733 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.921210 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.922654 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.923160 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.923893 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.924308 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:05 crc kubenswrapper[4733]: I0318 10:18:05.924342 4733 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 18 10:18:05 crc kubenswrapper[4733]: E0318 10:18:05.924617 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="200ms" Mar 18 10:18:06 crc kubenswrapper[4733]: E0318 10:18:06.125794 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="400ms" Mar 18 10:18:06 crc kubenswrapper[4733]: E0318 10:18:06.528370 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="800ms" Mar 18 10:18:07 crc kubenswrapper[4733]: E0318 10:18:07.329725 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="1.6s" Mar 18 10:18:07 crc kubenswrapper[4733]: E0318 10:18:07.634769 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.184:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189de8279c9a6215 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 10:18:05.082575381 +0000 UTC m=+324.574309706,LastTimestamp:2026-03-18 10:18:05.082575381 +0000 UTC m=+324.574309706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 10:18:08 crc kubenswrapper[4733]: E0318 10:18:08.931420 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="3.2s" Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.177988 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.422415 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" containerName="oauth-openshift" containerID="cri-o://2063ba38b8f338dff7686f6578cd42c9d0c532672eb45f293854b46ba18f0fea" gracePeriod=15 Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.584380 4733 generic.go:334] "Generic (PLEG): container finished" podID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" containerID="2063ba38b8f338dff7686f6578cd42c9d0c532672eb45f293854b46ba18f0fea" exitCode=0 Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.584452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" event={"ID":"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1","Type":"ContainerDied","Data":"2063ba38b8f338dff7686f6578cd42c9d0c532672eb45f293854b46ba18f0fea"} Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.934311 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.935136 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:11 crc kubenswrapper[4733]: I0318 10:18:11.935427 4733 status_manager.go:851] "Failed to get status for pod" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-n6hmz\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029070 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029122 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029149 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029208 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029227 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.029372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030091 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030109 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030206 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030344 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030384 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030404 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030417 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.030428 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131266 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131331 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131349 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131461 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131482 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131504 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4clx\" (UniqueName: \"kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131606 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") pod \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\" (UID: \"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1\") " Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.131972 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: E0318 10:18:12.132911 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.184:6443: connect: connection refused" interval="6.4s" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.138860 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.138984 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx" (OuterVolumeSpecName: "kube-api-access-h4clx") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "kube-api-access-h4clx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.139605 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.139903 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.140168 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.140378 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.140637 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.141031 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.142750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" (UID: "486eda8c-6e6f-4761-b28c-8aeb72fcfcc1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.175361 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.175952 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.176544 4733 status_manager.go:851] "Failed to get status for pod" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-n6hmz\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.190442 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.190478 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:12 crc kubenswrapper[4733]: E0318 10:18:12.190849 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.191347 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:12 crc kubenswrapper[4733]: W0318 10:18:12.207582 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-dc696d54957a20c5a5cfe41b4bb7d9a141673cb5d5e8a271f41febb11379b5d8 WatchSource:0}: Error finding container dc696d54957a20c5a5cfe41b4bb7d9a141673cb5d5e8a271f41febb11379b5d8: Status 404 returned error can't find the container with id dc696d54957a20c5a5cfe41b4bb7d9a141673cb5d5e8a271f41febb11379b5d8 Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232409 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232437 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232450 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232460 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232471 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232482 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232491 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232501 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4clx\" (UniqueName: \"kubernetes.io/projected/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-kube-api-access-h4clx\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.232510 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.595221 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.595229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" event={"ID":"486eda8c-6e6f-4761-b28c-8aeb72fcfcc1","Type":"ContainerDied","Data":"a4a546ed80545bf50a0d399d05bcd3718be5de86367b6c0e97b326427eeeb776"} Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.595821 4733 scope.go:117] "RemoveContainer" containerID="2063ba38b8f338dff7686f6578cd42c9d0c532672eb45f293854b46ba18f0fea" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.596568 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.596841 4733 status_manager.go:851] "Failed to get status for pod" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-n6hmz\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.599940 4733 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="caedb19efef737343ed2995bbe146d00288783b4e673a1c79a4ec21f417316dc" exitCode=0 Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.600015 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"caedb19efef737343ed2995bbe146d00288783b4e673a1c79a4ec21f417316dc"} Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.600064 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dc696d54957a20c5a5cfe41b4bb7d9a141673cb5d5e8a271f41febb11379b5d8"} Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.601258 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.601612 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.602280 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: E0318 10:18:12.602491 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.602910 4733 status_manager.go:851] "Failed to get status for pod" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-n6hmz\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.622209 4733 status_manager.go:851] "Failed to get status for pod" podUID="71bc6618-8df4-4a35-9469-772a853eff06" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:12 crc kubenswrapper[4733]: I0318 10:18:12.622792 4733 status_manager.go:851] "Failed to get status for pod" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" pod="openshift-authentication/oauth-openshift-558db77b4-n6hmz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-n6hmz\": dial tcp 38.102.83.184:6443: connect: connection refused" Mar 18 10:18:13 crc kubenswrapper[4733]: I0318 10:18:13.633055 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"101b56ee994525adbab0b16f8181f6df6c1e8c1a27b1f7f26f318c3625984c1f"} Mar 18 10:18:13 crc kubenswrapper[4733]: I0318 10:18:13.633426 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fc338a5575a4bed904037de792db05083f4715433f683a21c15af669a961168d"} Mar 18 10:18:13 crc kubenswrapper[4733]: I0318 10:18:13.633437 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c7441d26972d7d30232fee4709846b01e8d87f67f2beecda109a66454d99a988"} Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.643343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"961eee9018b09c4798c5cf5e9d4815d8c6df735727c5ebac8789e1399136eb39"} Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.643781 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.643797 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ad1cb73036760030ef554884132976824f996c9f56abba7b77c7033d32dfa23"} Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.643655 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.643824 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.645746 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.646313 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.646367 4733 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822" exitCode=1 Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.646393 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822"} Mar 18 10:18:14 crc kubenswrapper[4733]: I0318 10:18:14.646936 4733 scope.go:117] "RemoveContainer" containerID="7e84c65c99c9c698f4097bbffe0efebd320e4fc2c4a58788a606e7f0b98e1822" Mar 18 10:18:15 crc kubenswrapper[4733]: I0318 10:18:15.657636 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 10:18:15 crc kubenswrapper[4733]: I0318 10:18:15.658450 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 10:18:15 crc kubenswrapper[4733]: I0318 10:18:15.658500 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"569551dc2aa3d7b4b8162b0916c16e4ef346d8ca060ae26c5676163ad541f8f2"} Mar 18 10:18:16 crc kubenswrapper[4733]: I0318 10:18:16.643988 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:18:17 crc kubenswrapper[4733]: I0318 10:18:17.192091 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:17 crc kubenswrapper[4733]: I0318 10:18:17.192159 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:17 crc kubenswrapper[4733]: I0318 10:18:17.203350 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:18 crc kubenswrapper[4733]: I0318 10:18:18.889454 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:18:18 crc kubenswrapper[4733]: I0318 10:18:18.896623 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:18:19 crc kubenswrapper[4733]: I0318 10:18:19.659451 4733 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:19 crc kubenswrapper[4733]: I0318 10:18:19.688388 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:19 crc kubenswrapper[4733]: I0318 10:18:19.688439 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:19 crc kubenswrapper[4733]: I0318 10:18:19.696905 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:20 crc kubenswrapper[4733]: I0318 10:18:20.992260 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:20 crc kubenswrapper[4733]: I0318 10:18:20.992312 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ddb303e3-8922-4b43-9bba-2d3f0c30c6b8" Mar 18 10:18:21 crc kubenswrapper[4733]: I0318 10:18:21.190032 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="43b0fae6-f355-4c4a-a6d1-10662e33fa79" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.301121 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.301291 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.301376 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.301435 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.303409 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.304099 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.304116 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.304701 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.313479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.314020 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.321372 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.322303 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3650177-e338-4eba-ab42-bc0cd14c9d65-metrics-certs\") pod \"network-metrics-daemon-4s425\" (UID: \"b3650177-e338-4eba-ab42-bc0cd14c9d65\") " pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.327079 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.402842 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.408252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.498114 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.618140 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.620530 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.625896 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 10:18:23 crc kubenswrapper[4733]: I0318 10:18:23.628797 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4s425" Mar 18 10:18:24 crc kubenswrapper[4733]: W0318 10:18:24.178537 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-52c5d4fd6a6df793c0b80dca1e6d93c8b8b34acbd9d37159057b2efe0c813fb4 WatchSource:0}: Error finding container 52c5d4fd6a6df793c0b80dca1e6d93c8b8b34acbd9d37159057b2efe0c813fb4: Status 404 returned error can't find the container with id 52c5d4fd6a6df793c0b80dca1e6d93c8b8b34acbd9d37159057b2efe0c813fb4 Mar 18 10:18:24 crc kubenswrapper[4733]: W0318 10:18:24.256944 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3650177_e338_4eba_ab42_bc0cd14c9d65.slice/crio-03aba841ac2e1d2a33eb39d018365decea883523496253a386082ce0a0330775 WatchSource:0}: Error finding container 03aba841ac2e1d2a33eb39d018365decea883523496253a386082ce0a0330775: Status 404 returned error can't find the container with id 03aba841ac2e1d2a33eb39d018365decea883523496253a386082ce0a0330775 Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.024268 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ea66405bcfb3ac15e857ca56bfc2529bb805d39049cd5b8e1864cc65a95481b4"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.024812 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a09cb9227fb4162e36e81dce54e11238242d400b1d1035135a6683d0e4dc78cb"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.027249 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"269f3014e6f5c6075c765c0da6300d4eaf9926fd1df5ac0665716155c9042d44"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.027342 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8e97a2edc2af846924e5525c4644039e7d7375498d8a02fe8c5c3850861c5b45"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.030204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4s425" event={"ID":"b3650177-e338-4eba-ab42-bc0cd14c9d65","Type":"ContainerStarted","Data":"8e03bee5addf0ed44c0da8f14501e548939549f0389b5b4d987b33d193738139"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.030271 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4s425" event={"ID":"b3650177-e338-4eba-ab42-bc0cd14c9d65","Type":"ContainerStarted","Data":"ec7b75e20fa98c379acdef7414db972be5dfe4a9ed77cb9b8b2a54f4bd2dcaa8"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.030285 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4s425" event={"ID":"b3650177-e338-4eba-ab42-bc0cd14c9d65","Type":"ContainerStarted","Data":"03aba841ac2e1d2a33eb39d018365decea883523496253a386082ce0a0330775"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.032619 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c0afe4acc9ed318cef8afdacb201f65e8a1b8e6470bd18362461824eb194d68e"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.032676 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"52c5d4fd6a6df793c0b80dca1e6d93c8b8b34acbd9d37159057b2efe0c813fb4"} Mar 18 10:18:25 crc kubenswrapper[4733]: I0318 10:18:25.032835 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:18:26 crc kubenswrapper[4733]: I0318 10:18:26.043060 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Mar 18 10:18:26 crc kubenswrapper[4733]: I0318 10:18:26.043694 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="269f3014e6f5c6075c765c0da6300d4eaf9926fd1df5ac0665716155c9042d44" exitCode=255 Mar 18 10:18:26 crc kubenswrapper[4733]: I0318 10:18:26.043837 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"269f3014e6f5c6075c765c0da6300d4eaf9926fd1df5ac0665716155c9042d44"} Mar 18 10:18:26 crc kubenswrapper[4733]: I0318 10:18:26.044776 4733 scope.go:117] "RemoveContainer" containerID="269f3014e6f5c6075c765c0da6300d4eaf9926fd1df5ac0665716155c9042d44" Mar 18 10:18:26 crc kubenswrapper[4733]: I0318 10:18:26.653936 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 10:18:27 crc kubenswrapper[4733]: I0318 10:18:27.051542 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Mar 18 10:18:27 crc kubenswrapper[4733]: I0318 10:18:27.051589 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191"} Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.068300 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.069991 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.070059 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191" exitCode=255 Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.070127 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191"} Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.070274 4733 scope.go:117] "RemoveContainer" containerID="269f3014e6f5c6075c765c0da6300d4eaf9926fd1df5ac0665716155c9042d44" Mar 18 10:18:28 crc kubenswrapper[4733]: I0318 10:18:28.071228 4733 scope.go:117] "RemoveContainer" containerID="11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191" Mar 18 10:18:28 crc kubenswrapper[4733]: E0318 10:18:28.071623 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:18:29 crc kubenswrapper[4733]: I0318 10:18:29.082404 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Mar 18 10:18:30 crc kubenswrapper[4733]: I0318 10:18:30.529709 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 10:18:30 crc kubenswrapper[4733]: I0318 10:18:30.832243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 10:18:30 crc kubenswrapper[4733]: I0318 10:18:30.958843 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 10:18:31 crc kubenswrapper[4733]: I0318 10:18:31.180573 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 10:18:31 crc kubenswrapper[4733]: I0318 10:18:31.517984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 10:18:31 crc kubenswrapper[4733]: I0318 10:18:31.865573 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 10:18:31 crc kubenswrapper[4733]: I0318 10:18:31.927499 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 10:18:31 crc kubenswrapper[4733]: I0318 10:18:31.930019 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.021700 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.134915 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.156686 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.160496 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.185933 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.374080 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.541375 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.715143 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 10:18:32 crc kubenswrapper[4733]: I0318 10:18:32.772332 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.028013 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.203601 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.332705 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.578870 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.620229 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.860375 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.977564 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 10:18:33 crc kubenswrapper[4733]: I0318 10:18:33.995443 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.110442 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.202743 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.318319 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.348557 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.376402 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.405978 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.439275 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.521996 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.532960 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.539195 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.547525 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.553398 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.714408 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.786446 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.904291 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.921590 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.967336 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 10:18:34 crc kubenswrapper[4733]: I0318 10:18:34.973735 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.041123 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.056082 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.176882 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.607901 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.623826 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.639268 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.696160 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.715358 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.737552 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.744154 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.802046 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.872157 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.872168 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.958672 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 10:18:35 crc kubenswrapper[4733]: I0318 10:18:35.963101 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.087550 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.111913 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.128677 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.156338 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.190618 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.299315 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.303511 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.374492 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.507842 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.546801 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.551104 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.700883 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.723906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.740041 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.824941 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 10:18:36 crc kubenswrapper[4733]: I0318 10:18:36.955906 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.044809 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.085593 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.137530 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.150625 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.210574 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.238573 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.395655 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.434485 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.467392 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.500572 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.505290 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.575530 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.685850 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.694454 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.732543 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.832461 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 10:18:37 crc kubenswrapper[4733]: I0318 10:18:37.916247 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.001612 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.007074 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.211754 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.244004 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.267427 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.373252 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.379114 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.465156 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.484072 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.603312 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.640865 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.681335 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.702419 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.776882 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.904061 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.910028 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.940369 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.941822 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 10:18:38 crc kubenswrapper[4733]: I0318 10:18:38.945307 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.001722 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.105636 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.147602 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.202001 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.207360 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4s425" podStartSLOduration=301.207335519 podStartE2EDuration="5m1.207335519s" podCreationTimestamp="2026-03-18 10:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:18:25.103565048 +0000 UTC m=+344.595299403" watchObservedRunningTime="2026-03-18 10:18:39.207335519 +0000 UTC m=+358.699069874" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.209848 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-n6hmz"] Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.209937 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.209980 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4s425"] Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.215739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.246355 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.246326706 podStartE2EDuration="20.246326706s" podCreationTimestamp="2026-03-18 10:18:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:18:39.24013955 +0000 UTC m=+358.731873935" watchObservedRunningTime="2026-03-18 10:18:39.246326706 +0000 UTC m=+358.738061061" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.273398 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.286679 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.346034 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.347951 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.348617 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.384559 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.487000 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.513339 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.513876 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.702120 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.708519 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.743040 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 10:18:39 crc kubenswrapper[4733]: I0318 10:18:39.771687 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.034986 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.201690 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.221180 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.339630 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.427686 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.683634 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.729365 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.757403 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563818-44h4f"] Mar 18 10:18:40 crc kubenswrapper[4733]: E0318 10:18:40.757719 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" containerName="oauth-openshift" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.757743 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" containerName="oauth-openshift" Mar 18 10:18:40 crc kubenswrapper[4733]: E0318 10:18:40.757760 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bc6618-8df4-4a35-9469-772a853eff06" containerName="installer" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.757770 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bc6618-8df4-4a35-9469-772a853eff06" containerName="installer" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.757924 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="71bc6618-8df4-4a35-9469-772a853eff06" containerName="installer" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.757942 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" containerName="oauth-openshift" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.758686 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.765541 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563818-44h4f"] Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.767334 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.767338 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.767407 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.797675 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84w8q\" (UniqueName: \"kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q\") pod \"auto-csr-approver-29563818-44h4f\" (UID: \"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2\") " pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.849052 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.879763 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.898926 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84w8q\" (UniqueName: \"kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q\") pod \"auto-csr-approver-29563818-44h4f\" (UID: \"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2\") " pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.905869 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.925720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84w8q\" (UniqueName: \"kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q\") pod \"auto-csr-approver-29563818-44h4f\" (UID: \"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2\") " pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.974997 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 10:18:40 crc kubenswrapper[4733]: I0318 10:18:40.982388 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.083655 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.101112 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.115138 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.167023 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.183228 4733 scope.go:117] "RemoveContainer" containerID="11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.183917 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="486eda8c-6e6f-4761-b28c-8aeb72fcfcc1" path="/var/lib/kubelet/pods/486eda8c-6e6f-4761-b28c-8aeb72fcfcc1/volumes" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.220666 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.275630 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.377138 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.445093 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.447077 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.450744 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.472297 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.512906 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.531492 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.536704 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.550751 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563818-44h4f"] Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.556828 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.627009 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.747430 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.763463 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.794261 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.805030 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.826326 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.919590 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.946661 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.993831 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 10:18:41 crc kubenswrapper[4733]: I0318 10:18:41.994841 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.041946 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.119385 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.154814 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.180692 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.181040 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0"} Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.182672 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563818-44h4f" event={"ID":"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2","Type":"ContainerStarted","Data":"60980c1eaefeb0e6ad1eda08b1369c7dacc1d0cb042f4a7a9717e7a4226cf89a"} Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.231234 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.231504 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0" gracePeriod=5 Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.397167 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.595289 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.617974 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.696180 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.705917 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.760932 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.761584 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.842294 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.892841 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.930005 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 10:18:42 crc kubenswrapper[4733]: I0318 10:18:42.956332 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.033140 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.110361 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.143021 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.166918 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.195848 4733 generic.go:334] "Generic (PLEG): container finished" podID="c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" containerID="4287f6e7720d29c2928f6ce2bc4de5dd996378a83ad9d6dd58331a0b52048815" exitCode=0 Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.196613 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563818-44h4f" event={"ID":"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2","Type":"ContainerDied","Data":"4287f6e7720d29c2928f6ce2bc4de5dd996378a83ad9d6dd58331a0b52048815"} Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.198077 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.198962 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.199162 4733 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0" exitCode=255 Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.199215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0"} Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.199281 4733 scope.go:117] "RemoveContainer" containerID="11a8ded935f89e7b9d933f1c47e20a0e908f0e892416bc28ad676a8d61146191" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.212827 4733 scope.go:117] "RemoveContainer" containerID="7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0" Mar 18 10:18:43 crc kubenswrapper[4733]: E0318 10:18:43.213443 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.296491 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.353115 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.460422 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.477282 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.601678 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.612168 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.623985 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.702504 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.726556 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 10:18:43 crc kubenswrapper[4733]: I0318 10:18:43.992423 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.208884 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.223760 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.283458 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.373016 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.450380 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.494201 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.540440 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.618293 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.659140 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84w8q\" (UniqueName: \"kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q\") pod \"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2\" (UID: \"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2\") " Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.679571 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q" (OuterVolumeSpecName: "kube-api-access-84w8q") pod "c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" (UID: "c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2"). InnerVolumeSpecName "kube-api-access-84w8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.761092 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84w8q\" (UniqueName: \"kubernetes.io/projected/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2-kube-api-access-84w8q\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.800801 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.907006 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 10:18:44 crc kubenswrapper[4733]: I0318 10:18:44.979481 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.031918 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.089650 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.152467 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.217130 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563818-44h4f" event={"ID":"c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2","Type":"ContainerDied","Data":"60980c1eaefeb0e6ad1eda08b1369c7dacc1d0cb042f4a7a9717e7a4226cf89a"} Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.217204 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60980c1eaefeb0e6ad1eda08b1369c7dacc1d0cb042f4a7a9717e7a4226cf89a" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.221606 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563818-44h4f" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.287062 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.297753 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.368742 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.550697 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.557599 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.597225 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.649357 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.747645 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.841506 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 10:18:45 crc kubenswrapper[4733]: I0318 10:18:45.858291 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.012986 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.179331 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.223941 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.518586 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.530583 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.797762 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 10:18:46 crc kubenswrapper[4733]: I0318 10:18:46.940832 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.030023 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.057621 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.130994 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.148915 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.191559 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.221439 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.390841 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.498778 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.809553 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.809640 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.905846 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906278 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906658 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906832 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906066 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906325 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906769 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.906893 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.907564 4733 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.907686 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.907825 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.907945 4733 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:47 crc kubenswrapper[4733]: I0318 10:18:47.917953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.009034 4733 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.238106 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.238837 4733 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0" exitCode=137 Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.238936 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.238956 4733 scope.go:117] "RemoveContainer" containerID="2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.255928 4733 scope.go:117] "RemoveContainer" containerID="2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0" Mar 18 10:18:48 crc kubenswrapper[4733]: E0318 10:18:48.256587 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0\": container with ID starting with 2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0 not found: ID does not exist" containerID="2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.256637 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0"} err="failed to get container status \"2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0\": rpc error: code = NotFound desc = could not find container \"2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0\": container with ID starting with 2a3ba4486c7a6274fac695e45bc91d2e46d704c4c9018832a7072975d4fecee0 not found: ID does not exist" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.345902 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 10:18:48 crc kubenswrapper[4733]: I0318 10:18:48.814885 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.183470 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.383544 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.547268 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-s4zq9"] Mar 18 10:18:49 crc kubenswrapper[4733]: E0318 10:18:49.547534 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.547548 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 10:18:49 crc kubenswrapper[4733]: E0318 10:18:49.547567 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" containerName="oc" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.547574 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" containerName="oc" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.547695 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" containerName="oc" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.547710 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.548120 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.550887 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.551266 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.552020 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553113 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553272 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553324 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553475 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553244 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553663 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.553735 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.554747 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.555228 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.562333 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.567274 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-s4zq9"] Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.580943 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.583225 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628400 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628493 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628549 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khh9z\" (UniqueName: \"kubernetes.io/projected/2c695847-f9ab-4d8f-8e05-868fec637b86-kube-api-access-khh9z\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628621 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-policies\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628686 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-dir\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628741 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628773 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.628984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730770 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730797 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730824 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730850 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730885 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730904 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khh9z\" (UniqueName: \"kubernetes.io/projected/2c695847-f9ab-4d8f-8e05-868fec637b86-kube-api-access-khh9z\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730929 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-policies\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730955 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-dir\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.730996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.731026 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.731059 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.731117 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.731922 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.732141 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-dir\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.732668 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.733385 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-audit-policies\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.734637 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.736371 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.737085 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.737973 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.738529 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.739091 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.740584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.746877 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.756794 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2c695847-f9ab-4d8f-8e05-868fec637b86-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.756973 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khh9z\" (UniqueName: \"kubernetes.io/projected/2c695847-f9ab-4d8f-8e05-868fec637b86-kube-api-access-khh9z\") pod \"oauth-openshift-68755f559b-s4zq9\" (UID: \"2c695847-f9ab-4d8f-8e05-868fec637b86\") " pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:49 crc kubenswrapper[4733]: I0318 10:18:49.861882 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:50 crc kubenswrapper[4733]: I0318 10:18:50.300321 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-s4zq9"] Mar 18 10:18:50 crc kubenswrapper[4733]: W0318 10:18:50.312897 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c695847_f9ab_4d8f_8e05_868fec637b86.slice/crio-a8661530409955b5c494f8be4fcd86a397e0f4e70f27c610972ff3a633a2e60e WatchSource:0}: Error finding container a8661530409955b5c494f8be4fcd86a397e0f4e70f27c610972ff3a633a2e60e: Status 404 returned error can't find the container with id a8661530409955b5c494f8be4fcd86a397e0f4e70f27c610972ff3a633a2e60e Mar 18 10:18:51 crc kubenswrapper[4733]: I0318 10:18:51.271796 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" event={"ID":"2c695847-f9ab-4d8f-8e05-868fec637b86","Type":"ContainerStarted","Data":"4110f899e8265094fb2fbbd365d9ea24f37606fce5120f3e02a5bb0445cbe332"} Mar 18 10:18:51 crc kubenswrapper[4733]: I0318 10:18:51.272295 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" event={"ID":"2c695847-f9ab-4d8f-8e05-868fec637b86","Type":"ContainerStarted","Data":"a8661530409955b5c494f8be4fcd86a397e0f4e70f27c610972ff3a633a2e60e"} Mar 18 10:18:51 crc kubenswrapper[4733]: I0318 10:18:51.272340 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:51 crc kubenswrapper[4733]: I0318 10:18:51.281802 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" Mar 18 10:18:51 crc kubenswrapper[4733]: I0318 10:18:51.314945 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68755f559b-s4zq9" podStartSLOduration=65.314917027 podStartE2EDuration="1m5.314917027s" podCreationTimestamp="2026-03-18 10:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:18:51.313821686 +0000 UTC m=+370.805556021" watchObservedRunningTime="2026-03-18 10:18:51.314917027 +0000 UTC m=+370.806651372" Mar 18 10:18:57 crc kubenswrapper[4733]: I0318 10:18:57.181129 4733 scope.go:117] "RemoveContainer" containerID="7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0" Mar 18 10:18:57 crc kubenswrapper[4733]: E0318 10:18:57.182449 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 10:19:03 crc kubenswrapper[4733]: I0318 10:19:03.361777 4733 generic.go:334] "Generic (PLEG): container finished" podID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerID="3d92f9fbfa1c8b8490e331060d587d908cf420777497bf90bb4815f3f49e79dd" exitCode=0 Mar 18 10:19:03 crc kubenswrapper[4733]: I0318 10:19:03.361864 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerDied","Data":"3d92f9fbfa1c8b8490e331060d587d908cf420777497bf90bb4815f3f49e79dd"} Mar 18 10:19:03 crc kubenswrapper[4733]: I0318 10:19:03.363009 4733 scope.go:117] "RemoveContainer" containerID="3d92f9fbfa1c8b8490e331060d587d908cf420777497bf90bb4815f3f49e79dd" Mar 18 10:19:03 crc kubenswrapper[4733]: I0318 10:19:03.626157 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 10:19:04 crc kubenswrapper[4733]: I0318 10:19:04.369610 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerStarted","Data":"e99c56e1939c6c49ea2bc0d06c119ca0495ae09507c35a951b28f4145d07b5a2"} Mar 18 10:19:04 crc kubenswrapper[4733]: I0318 10:19:04.369980 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:19:04 crc kubenswrapper[4733]: I0318 10:19:04.371417 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:19:08 crc kubenswrapper[4733]: I0318 10:19:08.175828 4733 scope.go:117] "RemoveContainer" containerID="7157857ac224531b0050659a4ecc30a225cc12de5cd75c465a22ea80041c47d0" Mar 18 10:19:08 crc kubenswrapper[4733]: I0318 10:19:08.404011 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Mar 18 10:19:08 crc kubenswrapper[4733]: I0318 10:19:08.404502 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8f0aa49adcee1882dbf9476e69a98aecde2bd5d5ac8dc2661bf8de9a9e923a4a"} Mar 18 10:19:43 crc kubenswrapper[4733]: I0318 10:19:43.572093 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:19:43 crc kubenswrapper[4733]: I0318 10:19:43.572999 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.137968 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563820-x8mq8"] Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.139947 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.143976 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.145352 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.145841 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.154899 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563820-x8mq8"] Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.271871 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grxg2\" (UniqueName: \"kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2\") pod \"auto-csr-approver-29563820-x8mq8\" (UID: \"949d71ae-f754-4b5c-8c0b-fec8d374f27e\") " pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.373330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grxg2\" (UniqueName: \"kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2\") pod \"auto-csr-approver-29563820-x8mq8\" (UID: \"949d71ae-f754-4b5c-8c0b-fec8d374f27e\") " pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.402144 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grxg2\" (UniqueName: \"kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2\") pod \"auto-csr-approver-29563820-x8mq8\" (UID: \"949d71ae-f754-4b5c-8c0b-fec8d374f27e\") " pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.459566 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:00 crc kubenswrapper[4733]: I0318 10:20:00.946764 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563820-x8mq8"] Mar 18 10:20:01 crc kubenswrapper[4733]: I0318 10:20:01.774775 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" event={"ID":"949d71ae-f754-4b5c-8c0b-fec8d374f27e","Type":"ContainerStarted","Data":"651bd5a061b0713de78b4145e584b9b64603e782db5f257b9677273fe6364f7e"} Mar 18 10:20:02 crc kubenswrapper[4733]: I0318 10:20:02.787404 4733 generic.go:334] "Generic (PLEG): container finished" podID="949d71ae-f754-4b5c-8c0b-fec8d374f27e" containerID="4d1f85ec68f66c1e8dcc6134fd20cc9907c6036a83ddad6341fd815f0c10f145" exitCode=0 Mar 18 10:20:02 crc kubenswrapper[4733]: I0318 10:20:02.787497 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" event={"ID":"949d71ae-f754-4b5c-8c0b-fec8d374f27e","Type":"ContainerDied","Data":"4d1f85ec68f66c1e8dcc6134fd20cc9907c6036a83ddad6341fd815f0c10f145"} Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.783884 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mjkph"] Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.785871 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.803821 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mjkph"] Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933249 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933321 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-tls\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933369 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-bound-sa-token\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933430 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-certificates\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933897 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-trusted-ca\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.933991 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24zq9\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-kube-api-access-24zq9\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:03 crc kubenswrapper[4733]: I0318 10:20:03.969637 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.035949 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24zq9\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-kube-api-access-24zq9\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-tls\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036046 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036069 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-bound-sa-token\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036095 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-certificates\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036124 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.036153 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-trusted-ca\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.037651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.037765 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-trusted-ca\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.038808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-certificates\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.044233 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.044235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-registry-tls\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.068172 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-bound-sa-token\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.073925 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24zq9\" (UniqueName: \"kubernetes.io/projected/c0a31e7f-de02-4d48-8581-99a2fbf6a34f-kube-api-access-24zq9\") pod \"image-registry-66df7c8f76-mjkph\" (UID: \"c0a31e7f-de02-4d48-8581-99a2fbf6a34f\") " pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.104884 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.120808 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.238541 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grxg2\" (UniqueName: \"kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2\") pod \"949d71ae-f754-4b5c-8c0b-fec8d374f27e\" (UID: \"949d71ae-f754-4b5c-8c0b-fec8d374f27e\") " Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.242941 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2" (OuterVolumeSpecName: "kube-api-access-grxg2") pod "949d71ae-f754-4b5c-8c0b-fec8d374f27e" (UID: "949d71ae-f754-4b5c-8c0b-fec8d374f27e"). InnerVolumeSpecName "kube-api-access-grxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.340956 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grxg2\" (UniqueName: \"kubernetes.io/projected/949d71ae-f754-4b5c-8c0b-fec8d374f27e-kube-api-access-grxg2\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.352428 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mjkph"] Mar 18 10:20:04 crc kubenswrapper[4733]: W0318 10:20:04.356580 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0a31e7f_de02_4d48_8581_99a2fbf6a34f.slice/crio-7445fed31344d0de91797ecb9edaa7bed1664ab428aba817759b56ba37e6521f WatchSource:0}: Error finding container 7445fed31344d0de91797ecb9edaa7bed1664ab428aba817759b56ba37e6521f: Status 404 returned error can't find the container with id 7445fed31344d0de91797ecb9edaa7bed1664ab428aba817759b56ba37e6521f Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.803025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" event={"ID":"949d71ae-f754-4b5c-8c0b-fec8d374f27e","Type":"ContainerDied","Data":"651bd5a061b0713de78b4145e584b9b64603e782db5f257b9677273fe6364f7e"} Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.803639 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="651bd5a061b0713de78b4145e584b9b64603e782db5f257b9677273fe6364f7e" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.803083 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563820-x8mq8" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.807007 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" event={"ID":"c0a31e7f-de02-4d48-8581-99a2fbf6a34f","Type":"ContainerStarted","Data":"5fac677e322c63bc66da29cf2df8dccfe7c681a6387d8dcd8638f32db2639ad1"} Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.807727 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.807874 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" event={"ID":"c0a31e7f-de02-4d48-8581-99a2fbf6a34f","Type":"ContainerStarted","Data":"7445fed31344d0de91797ecb9edaa7bed1664ab428aba817759b56ba37e6521f"} Mar 18 10:20:04 crc kubenswrapper[4733]: I0318 10:20:04.832740 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" podStartSLOduration=1.8327092569999999 podStartE2EDuration="1.832709257s" podCreationTimestamp="2026-03-18 10:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:20:04.824418169 +0000 UTC m=+444.316152524" watchObservedRunningTime="2026-03-18 10:20:04.832709257 +0000 UTC m=+444.324443582" Mar 18 10:20:13 crc kubenswrapper[4733]: I0318 10:20:13.571250 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:20:13 crc kubenswrapper[4733]: I0318 10:20:13.572309 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:20:24 crc kubenswrapper[4733]: I0318 10:20:24.129486 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mjkph" Mar 18 10:20:24 crc kubenswrapper[4733]: I0318 10:20:24.246768 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.518744 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.520050 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rls2r" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="registry-server" containerID="cri-o://7e9cf80fc09f50439f722c47d01b38f7d154cd5514d553f0573a5303858564f5" gracePeriod=30 Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.536548 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.537035 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f92nl" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="registry-server" containerID="cri-o://d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b" gracePeriod=30 Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.545239 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.545546 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" containerID="cri-o://e99c56e1939c6c49ea2bc0d06c119ca0495ae09507c35a951b28f4145d07b5a2" gracePeriod=30 Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.553227 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.553566 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jb86w" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="registry-server" containerID="cri-o://36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014" gracePeriod=30 Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.567721 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z6qb2"] Mar 18 10:20:27 crc kubenswrapper[4733]: E0318 10:20:27.567977 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949d71ae-f754-4b5c-8c0b-fec8d374f27e" containerName="oc" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.567991 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="949d71ae-f754-4b5c-8c0b-fec8d374f27e" containerName="oc" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.568096 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="949d71ae-f754-4b5c-8c0b-fec8d374f27e" containerName="oc" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.568715 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.574065 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.574272 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrwxg" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" containerID="cri-o://33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" gracePeriod=30 Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.584253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z6qb2"] Mar 18 10:20:27 crc kubenswrapper[4733]: E0318 10:20:27.641432 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47 is running failed: container process not found" containerID="33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" cmd=["grpc_health_probe","-addr=:50051"] Mar 18 10:20:27 crc kubenswrapper[4733]: E0318 10:20:27.642686 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47 is running failed: container process not found" containerID="33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" cmd=["grpc_health_probe","-addr=:50051"] Mar 18 10:20:27 crc kubenswrapper[4733]: E0318 10:20:27.643280 4733 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47 is running failed: container process not found" containerID="33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" cmd=["grpc_health_probe","-addr=:50051"] Mar 18 10:20:27 crc kubenswrapper[4733]: E0318 10:20:27.643363 4733 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-hrwxg" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.653317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q486k\" (UniqueName: \"kubernetes.io/projected/8ae3847e-6357-46a1-9578-88deb6e1531b-kube-api-access-q486k\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.653366 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.653431 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.757036 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.757143 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q486k\" (UniqueName: \"kubernetes.io/projected/8ae3847e-6357-46a1-9578-88deb6e1531b-kube-api-access-q486k\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.757194 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.759190 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.766141 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8ae3847e-6357-46a1-9578-88deb6e1531b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.776328 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q486k\" (UniqueName: \"kubernetes.io/projected/8ae3847e-6357-46a1-9578-88deb6e1531b-kube-api-access-q486k\") pod \"marketplace-operator-79b997595-z6qb2\" (UID: \"8ae3847e-6357-46a1-9578-88deb6e1531b\") " pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.916411 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.975395 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:20:27 crc kubenswrapper[4733]: I0318 10:20:27.987383 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.012449 4733 generic.go:334] "Generic (PLEG): container finished" podID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerID="33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" exitCode=0 Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.013245 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerDied","Data":"33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.029760 4733 generic.go:334] "Generic (PLEG): container finished" podID="92996997-080b-42c9-bc2c-19c2e68db896" containerID="7e9cf80fc09f50439f722c47d01b38f7d154cd5514d553f0573a5303858564f5" exitCode=0 Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.029827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerDied","Data":"7e9cf80fc09f50439f722c47d01b38f7d154cd5514d553f0573a5303858564f5"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.031137 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.032068 4733 generic.go:334] "Generic (PLEG): container finished" podID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerID="36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014" exitCode=0 Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.032107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerDied","Data":"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.032126 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jb86w" event={"ID":"0fd306cb-05db-40e1-a1ec-9f811ce7fec0","Type":"ContainerDied","Data":"152bb2d9d2d5d61c127ef6162804e32f4f4e993fb3a1aa90d7238cb79aedf035"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.032146 4733 scope.go:117] "RemoveContainer" containerID="36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.032319 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jb86w" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.039783 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.044183 4733 generic.go:334] "Generic (PLEG): container finished" podID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerID="e99c56e1939c6c49ea2bc0d06c119ca0495ae09507c35a951b28f4145d07b5a2" exitCode=0 Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.044290 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerDied","Data":"e99c56e1939c6c49ea2bc0d06c119ca0495ae09507c35a951b28f4145d07b5a2"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.045563 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.051170 4733 generic.go:334] "Generic (PLEG): container finished" podID="527056ad-4daf-4dd5-9e31-887d55be0336" containerID="d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b" exitCode=0 Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.051211 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerDied","Data":"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.051230 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f92nl" event={"ID":"527056ad-4daf-4dd5-9e31-887d55be0336","Type":"ContainerDied","Data":"5d9e5dab0932c3cd3cd8b8f12fa8d0d49db59eddcefaa706bd16f11d86be1eac"} Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.051326 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f92nl" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063351 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j4k9\" (UniqueName: \"kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9\") pod \"527056ad-4daf-4dd5-9e31-887d55be0336\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063404 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content\") pod \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063483 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities\") pod \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content\") pod \"527056ad-4daf-4dd5-9e31-887d55be0336\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5frtf\" (UniqueName: \"kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf\") pod \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\" (UID: \"0fd306cb-05db-40e1-a1ec-9f811ce7fec0\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.063609 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities\") pod \"527056ad-4daf-4dd5-9e31-887d55be0336\" (UID: \"527056ad-4daf-4dd5-9e31-887d55be0336\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.066472 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities" (OuterVolumeSpecName: "utilities") pod "0fd306cb-05db-40e1-a1ec-9f811ce7fec0" (UID: "0fd306cb-05db-40e1-a1ec-9f811ce7fec0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.068176 4733 scope.go:117] "RemoveContainer" containerID="4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.070804 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf" (OuterVolumeSpecName: "kube-api-access-5frtf") pod "0fd306cb-05db-40e1-a1ec-9f811ce7fec0" (UID: "0fd306cb-05db-40e1-a1ec-9f811ce7fec0"). InnerVolumeSpecName "kube-api-access-5frtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.073125 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9" (OuterVolumeSpecName: "kube-api-access-9j4k9") pod "527056ad-4daf-4dd5-9e31-887d55be0336" (UID: "527056ad-4daf-4dd5-9e31-887d55be0336"). InnerVolumeSpecName "kube-api-access-9j4k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.078172 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities" (OuterVolumeSpecName: "utilities") pod "527056ad-4daf-4dd5-9e31-887d55be0336" (UID: "527056ad-4daf-4dd5-9e31-887d55be0336"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.120435 4733 scope.go:117] "RemoveContainer" containerID="deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.120629 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fd306cb-05db-40e1-a1ec-9f811ce7fec0" (UID: "0fd306cb-05db-40e1-a1ec-9f811ce7fec0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.139875 4733 scope.go:117] "RemoveContainer" containerID="36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.140537 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014\": container with ID starting with 36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014 not found: ID does not exist" containerID="36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.140586 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014"} err="failed to get container status \"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014\": rpc error: code = NotFound desc = could not find container \"36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014\": container with ID starting with 36d15214eccc522b73ee0fe4b5f5b4531b1d0593c4e73af5bdcac8f8e55d7014 not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.140623 4733 scope.go:117] "RemoveContainer" containerID="4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.141109 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9\": container with ID starting with 4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9 not found: ID does not exist" containerID="4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.141141 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9"} err="failed to get container status \"4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9\": rpc error: code = NotFound desc = could not find container \"4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9\": container with ID starting with 4da80ec2ba0c104ba8616114aa62d195906b3ceb35fe815aeee6c6a50ba00bd9 not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.141170 4733 scope.go:117] "RemoveContainer" containerID="deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.141529 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e\": container with ID starting with deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e not found: ID does not exist" containerID="deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.141587 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e"} err="failed to get container status \"deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e\": rpc error: code = NotFound desc = could not find container \"deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e\": container with ID starting with deb249a09e24f844e1f0eaad077e13c564da63c225d86fe92c3b3e169a3f2a0e not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.141627 4733 scope.go:117] "RemoveContainer" containerID="e99c56e1939c6c49ea2bc0d06c119ca0495ae09507c35a951b28f4145d07b5a2" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.153794 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "527056ad-4daf-4dd5-9e31-887d55be0336" (UID: "527056ad-4daf-4dd5-9e31-887d55be0336"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164534 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsw8d\" (UniqueName: \"kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d\") pod \"fb7ed879-1474-4200-88d4-70e425e2bcb1\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164601 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mscv\" (UniqueName: \"kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv\") pod \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164656 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities\") pod \"92996997-080b-42c9-bc2c-19c2e68db896\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164703 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities\") pod \"fb7ed879-1474-4200-88d4-70e425e2bcb1\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164726 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca\") pod \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164748 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content\") pod \"92996997-080b-42c9-bc2c-19c2e68db896\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164777 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7hv7\" (UniqueName: \"kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7\") pod \"92996997-080b-42c9-bc2c-19c2e68db896\" (UID: \"92996997-080b-42c9-bc2c-19c2e68db896\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164804 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content\") pod \"fb7ed879-1474-4200-88d4-70e425e2bcb1\" (UID: \"fb7ed879-1474-4200-88d4-70e425e2bcb1\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.164843 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics\") pod \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\" (UID: \"5192f67b-f2ab-45eb-9b1a-64bdff02437a\") " Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165111 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j4k9\" (UniqueName: \"kubernetes.io/projected/527056ad-4daf-4dd5-9e31-887d55be0336-kube-api-access-9j4k9\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165123 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165133 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165143 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165152 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5frtf\" (UniqueName: \"kubernetes.io/projected/0fd306cb-05db-40e1-a1ec-9f811ce7fec0-kube-api-access-5frtf\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.165162 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527056ad-4daf-4dd5-9e31-887d55be0336-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.169990 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5192f67b-f2ab-45eb-9b1a-64bdff02437a" (UID: "5192f67b-f2ab-45eb-9b1a-64bdff02437a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.170166 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities" (OuterVolumeSpecName: "utilities") pod "fb7ed879-1474-4200-88d4-70e425e2bcb1" (UID: "fb7ed879-1474-4200-88d4-70e425e2bcb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.171571 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5192f67b-f2ab-45eb-9b1a-64bdff02437a" (UID: "5192f67b-f2ab-45eb-9b1a-64bdff02437a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.172742 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities" (OuterVolumeSpecName: "utilities") pod "92996997-080b-42c9-bc2c-19c2e68db896" (UID: "92996997-080b-42c9-bc2c-19c2e68db896"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.172889 4733 scope.go:117] "RemoveContainer" containerID="3d92f9fbfa1c8b8490e331060d587d908cf420777497bf90bb4815f3f49e79dd" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.173304 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d" (OuterVolumeSpecName: "kube-api-access-jsw8d") pod "fb7ed879-1474-4200-88d4-70e425e2bcb1" (UID: "fb7ed879-1474-4200-88d4-70e425e2bcb1"). InnerVolumeSpecName "kube-api-access-jsw8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.175265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv" (OuterVolumeSpecName: "kube-api-access-8mscv") pod "5192f67b-f2ab-45eb-9b1a-64bdff02437a" (UID: "5192f67b-f2ab-45eb-9b1a-64bdff02437a"). InnerVolumeSpecName "kube-api-access-8mscv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.184710 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7" (OuterVolumeSpecName: "kube-api-access-w7hv7") pod "92996997-080b-42c9-bc2c-19c2e68db896" (UID: "92996997-080b-42c9-bc2c-19c2e68db896"). InnerVolumeSpecName "kube-api-access-w7hv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.193517 4733 scope.go:117] "RemoveContainer" containerID="d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.225599 4733 scope.go:117] "RemoveContainer" containerID="29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.241813 4733 scope.go:117] "RemoveContainer" containerID="83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.242975 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92996997-080b-42c9-bc2c-19c2e68db896" (UID: "92996997-080b-42c9-bc2c-19c2e68db896"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.257577 4733 scope.go:117] "RemoveContainer" containerID="d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.257971 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b\": container with ID starting with d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b not found: ID does not exist" containerID="d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.258043 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b"} err="failed to get container status \"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b\": rpc error: code = NotFound desc = could not find container \"d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b\": container with ID starting with d88a014dfa4a61b3bdf527747022f6d4b6201eb43fb9d2c08a1918862483878b not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.258079 4733 scope.go:117] "RemoveContainer" containerID="29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.258449 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4\": container with ID starting with 29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4 not found: ID does not exist" containerID="29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.258493 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4"} err="failed to get container status \"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4\": rpc error: code = NotFound desc = could not find container \"29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4\": container with ID starting with 29549f7b8e67a919f47e1ac510a621c5aca25e45afa1c1c52c5acdec0d566db4 not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.258525 4733 scope.go:117] "RemoveContainer" containerID="83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274" Mar 18 10:20:28 crc kubenswrapper[4733]: E0318 10:20:28.258835 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274\": container with ID starting with 83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274 not found: ID does not exist" containerID="83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.258862 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274"} err="failed to get container status \"83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274\": rpc error: code = NotFound desc = could not find container \"83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274\": container with ID starting with 83fe7a9d478dddba70a4985b321c90b2fd18ace1a534bec99183ab383ee3f274 not found: ID does not exist" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267022 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsw8d\" (UniqueName: \"kubernetes.io/projected/fb7ed879-1474-4200-88d4-70e425e2bcb1-kube-api-access-jsw8d\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267055 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mscv\" (UniqueName: \"kubernetes.io/projected/5192f67b-f2ab-45eb-9b1a-64bdff02437a-kube-api-access-8mscv\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267065 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267075 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267086 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267098 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92996997-080b-42c9-bc2c-19c2e68db896-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267107 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7hv7\" (UniqueName: \"kubernetes.io/projected/92996997-080b-42c9-bc2c-19c2e68db896-kube-api-access-w7hv7\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.267118 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5192f67b-f2ab-45eb-9b1a-64bdff02437a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.311224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb7ed879-1474-4200-88d4-70e425e2bcb1" (UID: "fb7ed879-1474-4200-88d4-70e425e2bcb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.368500 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb7ed879-1474-4200-88d4-70e425e2bcb1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.369001 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.372070 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jb86w"] Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.393073 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.396025 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f92nl"] Mar 18 10:20:28 crc kubenswrapper[4733]: I0318 10:20:28.425246 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z6qb2"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.059351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" event={"ID":"8ae3847e-6357-46a1-9578-88deb6e1531b","Type":"ContainerStarted","Data":"c456aa617e5a799611a817c36dd58fdd0a8f734f6e36771f134de22166e221d6"} Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.059433 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" event={"ID":"8ae3847e-6357-46a1-9578-88deb6e1531b","Type":"ContainerStarted","Data":"a9d8c901de60212a4bb21405b7e26db54d301aa6c882e75bd3bdd11b27f7bb2d"} Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.060074 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.063716 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrwxg" event={"ID":"fb7ed879-1474-4200-88d4-70e425e2bcb1","Type":"ContainerDied","Data":"62aa2aa87c6f58e0a138486db1e0ff0949ce50a5eef4891759673935a2791e3b"} Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.063758 4733 scope.go:117] "RemoveContainer" containerID="33010d46494372b311f8b2a190a49601d96469c4c865b75dc62dd08ddc447a47" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.063818 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrwxg" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.064240 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.068138 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rls2r" event={"ID":"92996997-080b-42c9-bc2c-19c2e68db896","Type":"ContainerDied","Data":"448a9e96bdf06f234c1da361f4be5cda2d36bf670a134ff4f206711028d80cac"} Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.068302 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rls2r" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.079062 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.079048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9h9xr" event={"ID":"5192f67b-f2ab-45eb-9b1a-64bdff02437a","Type":"ContainerDied","Data":"7e583c6a058ccd4e267ac556fbc1ecc397a1e062881c05b38f716c2d4a35947b"} Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.088016 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" podStartSLOduration=2.08798978 podStartE2EDuration="2.08798978s" podCreationTimestamp="2026-03-18 10:20:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:20:29.08236508 +0000 UTC m=+468.574099405" watchObservedRunningTime="2026-03-18 10:20:29.08798978 +0000 UTC m=+468.579724105" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.091539 4733 scope.go:117] "RemoveContainer" containerID="340cfa7d2b8654b1dea28355651bf6f54381a8104d827e6d38142ffcaf93e8ae" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.135250 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.146775 4733 scope.go:117] "RemoveContainer" containerID="aa0522bdc088c10a6b3c5dba1e3ad5057a62e8ded941287c75083cef63e55041" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.158152 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrwxg"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.172241 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.176829 4733 scope.go:117] "RemoveContainer" containerID="7e9cf80fc09f50439f722c47d01b38f7d154cd5514d553f0573a5303858564f5" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.196468 4733 scope.go:117] "RemoveContainer" containerID="9295312051c24cc07301903e63a22c698207253e2dd4d338c0be4c6fd4de6dec" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.226981 4733 scope.go:117] "RemoveContainer" containerID="a9bf744158dbc316b120322e1385bd5232386e738d2db0f1d91d2ac7d8a7ad1a" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.235283 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" path="/var/lib/kubelet/pods/0fd306cb-05db-40e1-a1ec-9f811ce7fec0/volumes" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.236306 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" path="/var/lib/kubelet/pods/527056ad-4daf-4dd5-9e31-887d55be0336/volumes" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.237005 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" path="/var/lib/kubelet/pods/fb7ed879-1474-4200-88d4-70e425e2bcb1/volumes" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.238195 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rls2r"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.238331 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kcbhw"] Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.238607 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.238676 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.238743 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.238797 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.238885 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.238947 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239018 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239072 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239133 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239189 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239368 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239442 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239533 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239592 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239648 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239702 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239765 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239824 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.239894 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.239958 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.240023 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.240087 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="extract-utilities" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.240168 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.240252 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.240368 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244421 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="extract-content" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244663 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244738 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fd306cb-05db-40e1-a1ec-9f811ce7fec0" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244832 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="92996997-080b-42c9-bc2c-19c2e68db896" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244900 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="527056ad-4daf-4dd5-9e31-887d55be0336" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.244967 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb7ed879-1474-4200-88d4-70e425e2bcb1" containerName="registry-server" Mar 18 10:20:29 crc kubenswrapper[4733]: E0318 10:20:29.245143 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.245231 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.245399 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" containerName="marketplace-operator" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.246141 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.246259 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9h9xr"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.246464 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcbhw"] Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.246520 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.248854 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.384348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-utilities\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.384438 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-catalog-content\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.384476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jzxm\" (UniqueName: \"kubernetes.io/projected/20640f37-bf35-4f24-abbb-b31cd00e5c9c-kube-api-access-5jzxm\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.487542 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-catalog-content\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.487650 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jzxm\" (UniqueName: \"kubernetes.io/projected/20640f37-bf35-4f24-abbb-b31cd00e5c9c-kube-api-access-5jzxm\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.487862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-utilities\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.488539 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-catalog-content\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.488747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20640f37-bf35-4f24-abbb-b31cd00e5c9c-utilities\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.524751 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jzxm\" (UniqueName: \"kubernetes.io/projected/20640f37-bf35-4f24-abbb-b31cd00e5c9c-kube-api-access-5jzxm\") pod \"redhat-marketplace-kcbhw\" (UID: \"20640f37-bf35-4f24-abbb-b31cd00e5c9c\") " pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.571141 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:29 crc kubenswrapper[4733]: I0318 10:20:29.832136 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kcbhw"] Mar 18 10:20:29 crc kubenswrapper[4733]: W0318 10:20:29.845358 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20640f37_bf35_4f24_abbb_b31cd00e5c9c.slice/crio-675eafbf68a5f68c2b8e076e71fb67c1665b94db59299648dbebe38a01d77c6e WatchSource:0}: Error finding container 675eafbf68a5f68c2b8e076e71fb67c1665b94db59299648dbebe38a01d77c6e: Status 404 returned error can't find the container with id 675eafbf68a5f68c2b8e076e71fb67c1665b94db59299648dbebe38a01d77c6e Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.096236 4733 generic.go:334] "Generic (PLEG): container finished" podID="20640f37-bf35-4f24-abbb-b31cd00e5c9c" containerID="a6279a4b196a2b80b7ac0476edb850e8a5114b94481124dfdd5dd9a1afe30a1a" exitCode=0 Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.096349 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcbhw" event={"ID":"20640f37-bf35-4f24-abbb-b31cd00e5c9c","Type":"ContainerDied","Data":"a6279a4b196a2b80b7ac0476edb850e8a5114b94481124dfdd5dd9a1afe30a1a"} Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.098585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcbhw" event={"ID":"20640f37-bf35-4f24-abbb-b31cd00e5c9c","Type":"ContainerStarted","Data":"675eafbf68a5f68c2b8e076e71fb67c1665b94db59299648dbebe38a01d77c6e"} Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.139525 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rs2b6"] Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.140926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.144959 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.155128 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs2b6"] Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.323028 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-utilities\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.323074 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-catalog-content\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.323358 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrr52\" (UniqueName: \"kubernetes.io/projected/3018dd18-ee9f-44a1-ab22-a6bddde19b31-kube-api-access-jrr52\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.424648 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-utilities\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.424696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-catalog-content\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.424758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrr52\" (UniqueName: \"kubernetes.io/projected/3018dd18-ee9f-44a1-ab22-a6bddde19b31-kube-api-access-jrr52\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.425741 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-utilities\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.426053 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3018dd18-ee9f-44a1-ab22-a6bddde19b31-catalog-content\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.446118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrr52\" (UniqueName: \"kubernetes.io/projected/3018dd18-ee9f-44a1-ab22-a6bddde19b31-kube-api-access-jrr52\") pod \"redhat-operators-rs2b6\" (UID: \"3018dd18-ee9f-44a1-ab22-a6bddde19b31\") " pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.467728 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:30 crc kubenswrapper[4733]: I0318 10:20:30.698978 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs2b6"] Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.125554 4733 generic.go:334] "Generic (PLEG): container finished" podID="20640f37-bf35-4f24-abbb-b31cd00e5c9c" containerID="174d079b92cada4004abb1338a1270574f6ce4e8b3c98eedfa1143baf90d2d39" exitCode=0 Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.125626 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcbhw" event={"ID":"20640f37-bf35-4f24-abbb-b31cd00e5c9c","Type":"ContainerDied","Data":"174d079b92cada4004abb1338a1270574f6ce4e8b3c98eedfa1143baf90d2d39"} Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.133428 4733 generic.go:334] "Generic (PLEG): container finished" podID="3018dd18-ee9f-44a1-ab22-a6bddde19b31" containerID="56f3163d6c13192463a53571cfbc21806b8894bc2c6d8bd6947c8fd7104dd2c3" exitCode=0 Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.134564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs2b6" event={"ID":"3018dd18-ee9f-44a1-ab22-a6bddde19b31","Type":"ContainerDied","Data":"56f3163d6c13192463a53571cfbc21806b8894bc2c6d8bd6947c8fd7104dd2c3"} Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.134649 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs2b6" event={"ID":"3018dd18-ee9f-44a1-ab22-a6bddde19b31","Type":"ContainerStarted","Data":"10566242ba9e1966bd518bca1f234dab4e3ed98028c6d262dba24228d30cb8d1"} Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.191937 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5192f67b-f2ab-45eb-9b1a-64bdff02437a" path="/var/lib/kubelet/pods/5192f67b-f2ab-45eb-9b1a-64bdff02437a/volumes" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.193001 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92996997-080b-42c9-bc2c-19c2e68db896" path="/var/lib/kubelet/pods/92996997-080b-42c9-bc2c-19c2e68db896/volumes" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.537447 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c9s27"] Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.539539 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.544615 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.551548 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9s27"] Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.695144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-utilities\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.695289 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxw6z\" (UniqueName: \"kubernetes.io/projected/a3f55919-82b3-4117-8734-cb9a26364d83-kube-api-access-jxw6z\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.695333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-catalog-content\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.797380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-utilities\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.797454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxw6z\" (UniqueName: \"kubernetes.io/projected/a3f55919-82b3-4117-8734-cb9a26364d83-kube-api-access-jxw6z\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.797490 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-catalog-content\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.798091 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-catalog-content\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.798090 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3f55919-82b3-4117-8734-cb9a26364d83-utilities\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.820667 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxw6z\" (UniqueName: \"kubernetes.io/projected/a3f55919-82b3-4117-8734-cb9a26364d83-kube-api-access-jxw6z\") pod \"certified-operators-c9s27\" (UID: \"a3f55919-82b3-4117-8734-cb9a26364d83\") " pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:31 crc kubenswrapper[4733]: I0318 10:20:31.925387 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.145831 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs2b6" event={"ID":"3018dd18-ee9f-44a1-ab22-a6bddde19b31","Type":"ContainerStarted","Data":"bd982dd2a6e8441453380cd896f110daacf18a1d8350c32f90c9d5ee2597cdc9"} Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.155890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kcbhw" event={"ID":"20640f37-bf35-4f24-abbb-b31cd00e5c9c","Type":"ContainerStarted","Data":"59625b60cde32a2df996e3f69224cf4c700e120ecd187c8f2bddd3a209e7c5d7"} Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.212850 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kcbhw" podStartSLOduration=1.7134267140000001 podStartE2EDuration="3.212821665s" podCreationTimestamp="2026-03-18 10:20:29 +0000 UTC" firstStartedPulling="2026-03-18 10:20:30.098387848 +0000 UTC m=+469.590122183" lastFinishedPulling="2026-03-18 10:20:31.597782799 +0000 UTC m=+471.089517134" observedRunningTime="2026-03-18 10:20:32.206863527 +0000 UTC m=+471.698597852" watchObservedRunningTime="2026-03-18 10:20:32.212821665 +0000 UTC m=+471.704555990" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.215028 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9s27"] Mar 18 10:20:32 crc kubenswrapper[4733]: W0318 10:20:32.224158 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3f55919_82b3_4117_8734_cb9a26364d83.slice/crio-efa271b227a89b0d5a1d77b150c9d7877c28dcc24ca0532ddf853caac8cbb6f9 WatchSource:0}: Error finding container efa271b227a89b0d5a1d77b150c9d7877c28dcc24ca0532ddf853caac8cbb6f9: Status 404 returned error can't find the container with id efa271b227a89b0d5a1d77b150c9d7877c28dcc24ca0532ddf853caac8cbb6f9 Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.534142 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-smkxx"] Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.535669 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.539251 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.545917 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smkxx"] Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.713729 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-utilities\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.713831 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lxn\" (UniqueName: \"kubernetes.io/projected/3b618f79-3791-49a8-a6aa-307fb25af727-kube-api-access-s8lxn\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.713866 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-catalog-content\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.816773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8lxn\" (UniqueName: \"kubernetes.io/projected/3b618f79-3791-49a8-a6aa-307fb25af727-kube-api-access-s8lxn\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.816843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-catalog-content\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.816895 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-utilities\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.817433 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-utilities\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.817626 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b618f79-3791-49a8-a6aa-307fb25af727-catalog-content\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:32 crc kubenswrapper[4733]: I0318 10:20:32.850570 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8lxn\" (UniqueName: \"kubernetes.io/projected/3b618f79-3791-49a8-a6aa-307fb25af727-kube-api-access-s8lxn\") pod \"community-operators-smkxx\" (UID: \"3b618f79-3791-49a8-a6aa-307fb25af727\") " pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.150503 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.181528 4733 generic.go:334] "Generic (PLEG): container finished" podID="a3f55919-82b3-4117-8734-cb9a26364d83" containerID="e606dfee6a97efc0a2e0b4fe618cac06e1e645ae870d8e8c6b065cc31aafe0d4" exitCode=0 Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.187776 4733 generic.go:334] "Generic (PLEG): container finished" podID="3018dd18-ee9f-44a1-ab22-a6bddde19b31" containerID="bd982dd2a6e8441453380cd896f110daacf18a1d8350c32f90c9d5ee2597cdc9" exitCode=0 Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.208429 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9s27" event={"ID":"a3f55919-82b3-4117-8734-cb9a26364d83","Type":"ContainerDied","Data":"e606dfee6a97efc0a2e0b4fe618cac06e1e645ae870d8e8c6b065cc31aafe0d4"} Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.208923 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9s27" event={"ID":"a3f55919-82b3-4117-8734-cb9a26364d83","Type":"ContainerStarted","Data":"efa271b227a89b0d5a1d77b150c9d7877c28dcc24ca0532ddf853caac8cbb6f9"} Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.208941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs2b6" event={"ID":"3018dd18-ee9f-44a1-ab22-a6bddde19b31","Type":"ContainerDied","Data":"bd982dd2a6e8441453380cd896f110daacf18a1d8350c32f90c9d5ee2597cdc9"} Mar 18 10:20:33 crc kubenswrapper[4733]: I0318 10:20:33.399421 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smkxx"] Mar 18 10:20:33 crc kubenswrapper[4733]: W0318 10:20:33.423642 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b618f79_3791_49a8_a6aa_307fb25af727.slice/crio-a1a400c266e61794ede357d7a38da9c390230097dfc9e30144046a9753588466 WatchSource:0}: Error finding container a1a400c266e61794ede357d7a38da9c390230097dfc9e30144046a9753588466: Status 404 returned error can't find the container with id a1a400c266e61794ede357d7a38da9c390230097dfc9e30144046a9753588466 Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.209268 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs2b6" event={"ID":"3018dd18-ee9f-44a1-ab22-a6bddde19b31","Type":"ContainerStarted","Data":"276b2767a0e52da4afffdbb245552b0f62633b7d2d1bbed274dbc46ec34e8057"} Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.211068 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9s27" event={"ID":"a3f55919-82b3-4117-8734-cb9a26364d83","Type":"ContainerStarted","Data":"4165647e230870e79d6ec6d3439c6520c6ee58ecbce291ff2fced2a81650c005"} Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.220514 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b618f79-3791-49a8-a6aa-307fb25af727" containerID="a6d690c2b27ccb116b3ec078eacfdb29e10df8cced9c0cc50fb4f8814235e950" exitCode=0 Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.220577 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smkxx" event={"ID":"3b618f79-3791-49a8-a6aa-307fb25af727","Type":"ContainerDied","Data":"a6d690c2b27ccb116b3ec078eacfdb29e10df8cced9c0cc50fb4f8814235e950"} Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.220613 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smkxx" event={"ID":"3b618f79-3791-49a8-a6aa-307fb25af727","Type":"ContainerStarted","Data":"a1a400c266e61794ede357d7a38da9c390230097dfc9e30144046a9753588466"} Mar 18 10:20:34 crc kubenswrapper[4733]: I0318 10:20:34.231953 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rs2b6" podStartSLOduration=1.748036549 podStartE2EDuration="4.231938735s" podCreationTimestamp="2026-03-18 10:20:30 +0000 UTC" firstStartedPulling="2026-03-18 10:20:31.13538435 +0000 UTC m=+470.627118715" lastFinishedPulling="2026-03-18 10:20:33.619286576 +0000 UTC m=+473.111020901" observedRunningTime="2026-03-18 10:20:34.230929207 +0000 UTC m=+473.722663532" watchObservedRunningTime="2026-03-18 10:20:34.231938735 +0000 UTC m=+473.723673080" Mar 18 10:20:35 crc kubenswrapper[4733]: I0318 10:20:35.229419 4733 generic.go:334] "Generic (PLEG): container finished" podID="a3f55919-82b3-4117-8734-cb9a26364d83" containerID="4165647e230870e79d6ec6d3439c6520c6ee58ecbce291ff2fced2a81650c005" exitCode=0 Mar 18 10:20:35 crc kubenswrapper[4733]: I0318 10:20:35.229498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9s27" event={"ID":"a3f55919-82b3-4117-8734-cb9a26364d83","Type":"ContainerDied","Data":"4165647e230870e79d6ec6d3439c6520c6ee58ecbce291ff2fced2a81650c005"} Mar 18 10:20:36 crc kubenswrapper[4733]: I0318 10:20:36.240642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9s27" event={"ID":"a3f55919-82b3-4117-8734-cb9a26364d83","Type":"ContainerStarted","Data":"233210e901efd5f8a61568c3871e0630b9a0f5d9872050ea82660e858137b2ee"} Mar 18 10:20:36 crc kubenswrapper[4733]: I0318 10:20:36.243394 4733 generic.go:334] "Generic (PLEG): container finished" podID="3b618f79-3791-49a8-a6aa-307fb25af727" containerID="4242ac68926bca06f761881b3ad94794f94a8733f43e8e77ca87a194635b909f" exitCode=0 Mar 18 10:20:36 crc kubenswrapper[4733]: I0318 10:20:36.243451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smkxx" event={"ID":"3b618f79-3791-49a8-a6aa-307fb25af727","Type":"ContainerDied","Data":"4242ac68926bca06f761881b3ad94794f94a8733f43e8e77ca87a194635b909f"} Mar 18 10:20:36 crc kubenswrapper[4733]: I0318 10:20:36.271721 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c9s27" podStartSLOduration=2.844417524 podStartE2EDuration="5.27169448s" podCreationTimestamp="2026-03-18 10:20:31 +0000 UTC" firstStartedPulling="2026-03-18 10:20:33.183919622 +0000 UTC m=+472.675653957" lastFinishedPulling="2026-03-18 10:20:35.611196588 +0000 UTC m=+475.102930913" observedRunningTime="2026-03-18 10:20:36.268808919 +0000 UTC m=+475.760543244" watchObservedRunningTime="2026-03-18 10:20:36.27169448 +0000 UTC m=+475.763428805" Mar 18 10:20:37 crc kubenswrapper[4733]: I0318 10:20:37.253630 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smkxx" event={"ID":"3b618f79-3791-49a8-a6aa-307fb25af727","Type":"ContainerStarted","Data":"de1f82b2a6315c18123f4ab94fb2f046135a4829f1ab2eff1b0c278227e5cd61"} Mar 18 10:20:39 crc kubenswrapper[4733]: I0318 10:20:39.571389 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:39 crc kubenswrapper[4733]: I0318 10:20:39.571598 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:39 crc kubenswrapper[4733]: I0318 10:20:39.637313 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:39 crc kubenswrapper[4733]: I0318 10:20:39.657672 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-smkxx" podStartSLOduration=5.109467928 podStartE2EDuration="7.657647102s" podCreationTimestamp="2026-03-18 10:20:32 +0000 UTC" firstStartedPulling="2026-03-18 10:20:34.222296183 +0000 UTC m=+473.714030518" lastFinishedPulling="2026-03-18 10:20:36.770475367 +0000 UTC m=+476.262209692" observedRunningTime="2026-03-18 10:20:37.289654583 +0000 UTC m=+476.781388908" watchObservedRunningTime="2026-03-18 10:20:39.657647102 +0000 UTC m=+479.149381427" Mar 18 10:20:40 crc kubenswrapper[4733]: I0318 10:20:40.362227 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kcbhw" Mar 18 10:20:40 crc kubenswrapper[4733]: I0318 10:20:40.469070 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:40 crc kubenswrapper[4733]: I0318 10:20:40.469132 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:41 crc kubenswrapper[4733]: I0318 10:20:41.525485 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rs2b6" podUID="3018dd18-ee9f-44a1-ab22-a6bddde19b31" containerName="registry-server" probeResult="failure" output=< Mar 18 10:20:41 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:20:41 crc kubenswrapper[4733]: > Mar 18 10:20:41 crc kubenswrapper[4733]: I0318 10:20:41.926440 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:41 crc kubenswrapper[4733]: I0318 10:20:41.926896 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:41 crc kubenswrapper[4733]: I0318 10:20:41.991967 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:42 crc kubenswrapper[4733]: I0318 10:20:42.363302 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c9s27" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.150996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.151059 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.244522 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.364345 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-smkxx" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.571487 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.571584 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.571652 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.572824 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:20:43 crc kubenswrapper[4733]: I0318 10:20:43.572932 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028" gracePeriod=600 Mar 18 10:20:44 crc kubenswrapper[4733]: I0318 10:20:44.337573 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028" exitCode=0 Mar 18 10:20:44 crc kubenswrapper[4733]: I0318 10:20:44.337736 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028"} Mar 18 10:20:44 crc kubenswrapper[4733]: I0318 10:20:44.338457 4733 scope.go:117] "RemoveContainer" containerID="615e7a90421535b4f8ff5e3b3a0ad9c958710094ffa4e3e4eb3eb41c79f80830" Mar 18 10:20:45 crc kubenswrapper[4733]: I0318 10:20:45.346897 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527"} Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.306338 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" podUID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" containerName="registry" containerID="cri-o://d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085" gracePeriod=30 Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.683115 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.838958 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qqrn\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839016 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839316 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839399 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839437 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.839524 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates\") pod \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\" (UID: \"7b5dc098-4a15-429b-8243-1ac75ce2e0c1\") " Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.840722 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.844034 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.851782 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.854876 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn" (OuterVolumeSpecName: "kube-api-access-2qqrn") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "kube-api-access-2qqrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.855397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.855652 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.859161 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.870886 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7b5dc098-4a15-429b-8243-1ac75ce2e0c1" (UID: "7b5dc098-4a15-429b-8243-1ac75ce2e0c1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.941789 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qqrn\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-kube-api-access-2qqrn\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.942238 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.942345 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.942420 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.942502 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.943021 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:49 crc kubenswrapper[4733]: I0318 10:20:49.943116 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7b5dc098-4a15-429b-8243-1ac75ce2e0c1-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.405136 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" containerID="d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085" exitCode=0 Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.405219 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.405219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" event={"ID":"7b5dc098-4a15-429b-8243-1ac75ce2e0c1","Type":"ContainerDied","Data":"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085"} Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.405354 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nwhtg" event={"ID":"7b5dc098-4a15-429b-8243-1ac75ce2e0c1","Type":"ContainerDied","Data":"bf9beab436bdff3f99c6c06c629fb5de1f2bcd079250aacd7d55627140dc6e11"} Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.405380 4733 scope.go:117] "RemoveContainer" containerID="d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.441761 4733 scope.go:117] "RemoveContainer" containerID="d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085" Mar 18 10:20:50 crc kubenswrapper[4733]: E0318 10:20:50.442645 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085\": container with ID starting with d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085 not found: ID does not exist" containerID="d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.442893 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085"} err="failed to get container status \"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085\": rpc error: code = NotFound desc = could not find container \"d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085\": container with ID starting with d196ee7bec70e95ed9ff3308e0424855deaa072d0c4faba514ae98e1dcaec085 not found: ID does not exist" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.463749 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.470676 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nwhtg"] Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.538470 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:50 crc kubenswrapper[4733]: I0318 10:20:50.599578 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rs2b6" Mar 18 10:20:51 crc kubenswrapper[4733]: I0318 10:20:51.190226 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" path="/var/lib/kubelet/pods/7b5dc098-4a15-429b-8243-1ac75ce2e0c1/volumes" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.152981 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563822-4fvb8"] Mar 18 10:22:00 crc kubenswrapper[4733]: E0318 10:22:00.154668 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" containerName="registry" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.154697 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" containerName="registry" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.155609 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5dc098-4a15-429b-8243-1ac75ce2e0c1" containerName="registry" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.156301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.159987 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.160380 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.160693 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.165797 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563822-4fvb8"] Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.295529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whsqs\" (UniqueName: \"kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs\") pod \"auto-csr-approver-29563822-4fvb8\" (UID: \"3f93d05a-41f2-4422-88aa-9dfddb13191f\") " pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.397877 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whsqs\" (UniqueName: \"kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs\") pod \"auto-csr-approver-29563822-4fvb8\" (UID: \"3f93d05a-41f2-4422-88aa-9dfddb13191f\") " pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.428803 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whsqs\" (UniqueName: \"kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs\") pod \"auto-csr-approver-29563822-4fvb8\" (UID: \"3f93d05a-41f2-4422-88aa-9dfddb13191f\") " pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:00 crc kubenswrapper[4733]: I0318 10:22:00.493427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:01 crc kubenswrapper[4733]: I0318 10:22:01.393176 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563822-4fvb8"] Mar 18 10:22:01 crc kubenswrapper[4733]: W0318 10:22:01.408684 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f93d05a_41f2_4422_88aa_9dfddb13191f.slice/crio-31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa WatchSource:0}: Error finding container 31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa: Status 404 returned error can't find the container with id 31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa Mar 18 10:22:01 crc kubenswrapper[4733]: I0318 10:22:01.413558 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:22:01 crc kubenswrapper[4733]: I0318 10:22:01.996750 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" event={"ID":"3f93d05a-41f2-4422-88aa-9dfddb13191f","Type":"ContainerStarted","Data":"31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa"} Mar 18 10:22:03 crc kubenswrapper[4733]: I0318 10:22:03.007591 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f93d05a-41f2-4422-88aa-9dfddb13191f" containerID="36c7a80bc1a34092c9183dbd958b5c05ea904377be8cffacb7112a1b4663e6a6" exitCode=0 Mar 18 10:22:03 crc kubenswrapper[4733]: I0318 10:22:03.007910 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" event={"ID":"3f93d05a-41f2-4422-88aa-9dfddb13191f","Type":"ContainerDied","Data":"36c7a80bc1a34092c9183dbd958b5c05ea904377be8cffacb7112a1b4663e6a6"} Mar 18 10:22:04 crc kubenswrapper[4733]: I0318 10:22:04.338771 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:04 crc kubenswrapper[4733]: I0318 10:22:04.458751 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whsqs\" (UniqueName: \"kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs\") pod \"3f93d05a-41f2-4422-88aa-9dfddb13191f\" (UID: \"3f93d05a-41f2-4422-88aa-9dfddb13191f\") " Mar 18 10:22:04 crc kubenswrapper[4733]: I0318 10:22:04.466627 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs" (OuterVolumeSpecName: "kube-api-access-whsqs") pod "3f93d05a-41f2-4422-88aa-9dfddb13191f" (UID: "3f93d05a-41f2-4422-88aa-9dfddb13191f"). InnerVolumeSpecName "kube-api-access-whsqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:22:04 crc kubenswrapper[4733]: I0318 10:22:04.560887 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whsqs\" (UniqueName: \"kubernetes.io/projected/3f93d05a-41f2-4422-88aa-9dfddb13191f-kube-api-access-whsqs\") on node \"crc\" DevicePath \"\"" Mar 18 10:22:05 crc kubenswrapper[4733]: I0318 10:22:05.025181 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" event={"ID":"3f93d05a-41f2-4422-88aa-9dfddb13191f","Type":"ContainerDied","Data":"31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa"} Mar 18 10:22:05 crc kubenswrapper[4733]: I0318 10:22:05.025299 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563822-4fvb8" Mar 18 10:22:05 crc kubenswrapper[4733]: I0318 10:22:05.025312 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31735bf21c4d4370dc615feda0ad1646f4d6cdd954f38a88ca325149075006fa" Mar 18 10:22:05 crc kubenswrapper[4733]: I0318 10:22:05.421666 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563816-4582s"] Mar 18 10:22:05 crc kubenswrapper[4733]: I0318 10:22:05.428119 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563816-4582s"] Mar 18 10:22:07 crc kubenswrapper[4733]: I0318 10:22:07.186130 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71a70c3c-d483-43f4-9f54-10978c7f8cc8" path="/var/lib/kubelet/pods/71a70c3c-d483-43f4-9f54-10978c7f8cc8/volumes" Mar 18 10:23:13 crc kubenswrapper[4733]: I0318 10:23:13.571771 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:23:13 crc kubenswrapper[4733]: I0318 10:23:13.572824 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:23:43 crc kubenswrapper[4733]: I0318 10:23:43.571734 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:23:43 crc kubenswrapper[4733]: I0318 10:23:43.572489 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:23:54 crc kubenswrapper[4733]: I0318 10:23:54.538673 4733 scope.go:117] "RemoveContainer" containerID="869578488a5526adb52c0d5efeb676ea68e5c20e95b1cf2d208fa00dbd02baca" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.152033 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563824-l49xk"] Mar 18 10:24:00 crc kubenswrapper[4733]: E0318 10:24:00.153450 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f93d05a-41f2-4422-88aa-9dfddb13191f" containerName="oc" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.153483 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f93d05a-41f2-4422-88aa-9dfddb13191f" containerName="oc" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.153723 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f93d05a-41f2-4422-88aa-9dfddb13191f" containerName="oc" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.154642 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.158745 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.159285 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.159411 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.167403 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563824-l49xk"] Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.325655 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89gfz\" (UniqueName: \"kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz\") pod \"auto-csr-approver-29563824-l49xk\" (UID: \"6077c15f-e285-4625-b336-a84327b1af2d\") " pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.427508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89gfz\" (UniqueName: \"kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz\") pod \"auto-csr-approver-29563824-l49xk\" (UID: \"6077c15f-e285-4625-b336-a84327b1af2d\") " pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.454923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89gfz\" (UniqueName: \"kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz\") pod \"auto-csr-approver-29563824-l49xk\" (UID: \"6077c15f-e285-4625-b336-a84327b1af2d\") " pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.484336 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.711674 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563824-l49xk"] Mar 18 10:24:00 crc kubenswrapper[4733]: I0318 10:24:00.869961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563824-l49xk" event={"ID":"6077c15f-e285-4625-b336-a84327b1af2d","Type":"ContainerStarted","Data":"6f54837142a70aebab3aa9bcb8b9b38eb2ba43f3e59b7b9ba5aa264c07ff2706"} Mar 18 10:24:01 crc kubenswrapper[4733]: I0318 10:24:01.880101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563824-l49xk" event={"ID":"6077c15f-e285-4625-b336-a84327b1af2d","Type":"ContainerStarted","Data":"6f50555f9faf96f94c8c33f53803364eb9620cbe1dd5e27e68cba9056a299fa1"} Mar 18 10:24:01 crc kubenswrapper[4733]: I0318 10:24:01.894116 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563824-l49xk" podStartSLOduration=1.116388676 podStartE2EDuration="1.89409588s" podCreationTimestamp="2026-03-18 10:24:00 +0000 UTC" firstStartedPulling="2026-03-18 10:24:00.713224348 +0000 UTC m=+680.204958683" lastFinishedPulling="2026-03-18 10:24:01.490931542 +0000 UTC m=+680.982665887" observedRunningTime="2026-03-18 10:24:01.891578329 +0000 UTC m=+681.383312654" watchObservedRunningTime="2026-03-18 10:24:01.89409588 +0000 UTC m=+681.385830205" Mar 18 10:24:02 crc kubenswrapper[4733]: I0318 10:24:02.891435 4733 generic.go:334] "Generic (PLEG): container finished" podID="6077c15f-e285-4625-b336-a84327b1af2d" containerID="6f50555f9faf96f94c8c33f53803364eb9620cbe1dd5e27e68cba9056a299fa1" exitCode=0 Mar 18 10:24:02 crc kubenswrapper[4733]: I0318 10:24:02.891507 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563824-l49xk" event={"ID":"6077c15f-e285-4625-b336-a84327b1af2d","Type":"ContainerDied","Data":"6f50555f9faf96f94c8c33f53803364eb9620cbe1dd5e27e68cba9056a299fa1"} Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.152516 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.270829 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563818-44h4f"] Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.276980 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563818-44h4f"] Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.281493 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89gfz\" (UniqueName: \"kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz\") pod \"6077c15f-e285-4625-b336-a84327b1af2d\" (UID: \"6077c15f-e285-4625-b336-a84327b1af2d\") " Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.290469 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz" (OuterVolumeSpecName: "kube-api-access-89gfz") pod "6077c15f-e285-4625-b336-a84327b1af2d" (UID: "6077c15f-e285-4625-b336-a84327b1af2d"). InnerVolumeSpecName "kube-api-access-89gfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.383837 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89gfz\" (UniqueName: \"kubernetes.io/projected/6077c15f-e285-4625-b336-a84327b1af2d-kube-api-access-89gfz\") on node \"crc\" DevicePath \"\"" Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.908005 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563824-l49xk" event={"ID":"6077c15f-e285-4625-b336-a84327b1af2d","Type":"ContainerDied","Data":"6f54837142a70aebab3aa9bcb8b9b38eb2ba43f3e59b7b9ba5aa264c07ff2706"} Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.908068 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f54837142a70aebab3aa9bcb8b9b38eb2ba43f3e59b7b9ba5aa264c07ff2706" Mar 18 10:24:04 crc kubenswrapper[4733]: I0318 10:24:04.908140 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563824-l49xk" Mar 18 10:24:05 crc kubenswrapper[4733]: I0318 10:24:05.191150 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2" path="/var/lib/kubelet/pods/c7d7efa6-dd10-4ee1-a93b-13ae5f74ebe2/volumes" Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.571953 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.574365 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.574580 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.575538 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.575773 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527" gracePeriod=600 Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.982947 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527" exitCode=0 Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.983144 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527"} Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.983429 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c"} Mar 18 10:24:13 crc kubenswrapper[4733]: I0318 10:24:13.983461 4733 scope.go:117] "RemoveContainer" containerID="2dcc5035fa17fe3e92cf26ce37e02cacce4ad31a0593e6e1184b98062f31f028" Mar 18 10:24:54 crc kubenswrapper[4733]: I0318 10:24:54.614360 4733 scope.go:117] "RemoveContainer" containerID="4287f6e7720d29c2928f6ce2bc4de5dd996378a83ad9d6dd58331a0b52048815" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.129242 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-8ds68"] Mar 18 10:25:35 crc kubenswrapper[4733]: E0318 10:25:35.130256 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6077c15f-e285-4625-b336-a84327b1af2d" containerName="oc" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.130275 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6077c15f-e285-4625-b336-a84327b1af2d" containerName="oc" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.130405 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6077c15f-e285-4625-b336-a84327b1af2d" containerName="oc" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.130834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.136428 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rjbj6" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.136506 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.143139 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-rd2dh"] Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.144886 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-rd2dh" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.152793 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.153635 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jd7pl" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.166849 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-8ds68"] Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.173818 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-2nr27"] Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.177973 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.183545 4733 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mcqkw" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.196415 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-rd2dh"] Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.196488 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-2nr27"] Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.228948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq9p9\" (UniqueName: \"kubernetes.io/projected/534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8-kube-api-access-cq9p9\") pod \"cert-manager-webhook-687f57d79b-2nr27\" (UID: \"534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8\") " pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.331281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq9p9\" (UniqueName: \"kubernetes.io/projected/534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8-kube-api-access-cq9p9\") pod \"cert-manager-webhook-687f57d79b-2nr27\" (UID: \"534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8\") " pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.331432 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9q6g\" (UniqueName: \"kubernetes.io/projected/ce77d29d-b82e-46be-a694-b6eea5da9379-kube-api-access-f9q6g\") pod \"cert-manager-858654f9db-rd2dh\" (UID: \"ce77d29d-b82e-46be-a694-b6eea5da9379\") " pod="cert-manager/cert-manager-858654f9db-rd2dh" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.331528 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8x85\" (UniqueName: \"kubernetes.io/projected/585c06be-86bd-48b7-954e-9aec01b08874-kube-api-access-s8x85\") pod \"cert-manager-cainjector-cf98fcc89-8ds68\" (UID: \"585c06be-86bd-48b7-954e-9aec01b08874\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.360545 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq9p9\" (UniqueName: \"kubernetes.io/projected/534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8-kube-api-access-cq9p9\") pod \"cert-manager-webhook-687f57d79b-2nr27\" (UID: \"534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8\") " pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.432856 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9q6g\" (UniqueName: \"kubernetes.io/projected/ce77d29d-b82e-46be-a694-b6eea5da9379-kube-api-access-f9q6g\") pod \"cert-manager-858654f9db-rd2dh\" (UID: \"ce77d29d-b82e-46be-a694-b6eea5da9379\") " pod="cert-manager/cert-manager-858654f9db-rd2dh" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.432947 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8x85\" (UniqueName: \"kubernetes.io/projected/585c06be-86bd-48b7-954e-9aec01b08874-kube-api-access-s8x85\") pod \"cert-manager-cainjector-cf98fcc89-8ds68\" (UID: \"585c06be-86bd-48b7-954e-9aec01b08874\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.450602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9q6g\" (UniqueName: \"kubernetes.io/projected/ce77d29d-b82e-46be-a694-b6eea5da9379-kube-api-access-f9q6g\") pod \"cert-manager-858654f9db-rd2dh\" (UID: \"ce77d29d-b82e-46be-a694-b6eea5da9379\") " pod="cert-manager/cert-manager-858654f9db-rd2dh" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.463532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8x85\" (UniqueName: \"kubernetes.io/projected/585c06be-86bd-48b7-954e-9aec01b08874-kube-api-access-s8x85\") pod \"cert-manager-cainjector-cf98fcc89-8ds68\" (UID: \"585c06be-86bd-48b7-954e-9aec01b08874\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.474299 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.486180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-rd2dh" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.499749 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:35 crc kubenswrapper[4733]: I0318 10:25:35.761307 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-rd2dh"] Mar 18 10:25:36 crc kubenswrapper[4733]: I0318 10:25:36.014047 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-8ds68"] Mar 18 10:25:36 crc kubenswrapper[4733]: W0318 10:25:36.019366 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod585c06be_86bd_48b7_954e_9aec01b08874.slice/crio-5c8591c08d39a2c1dcf39bc2b94cf19e43b4ae811739fc182daf63959295a448 WatchSource:0}: Error finding container 5c8591c08d39a2c1dcf39bc2b94cf19e43b4ae811739fc182daf63959295a448: Status 404 returned error can't find the container with id 5c8591c08d39a2c1dcf39bc2b94cf19e43b4ae811739fc182daf63959295a448 Mar 18 10:25:36 crc kubenswrapper[4733]: I0318 10:25:36.019743 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-2nr27"] Mar 18 10:25:36 crc kubenswrapper[4733]: W0318 10:25:36.031909 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod534b0ac6_c9b1_4940_9e6e_ed36de1ec1e8.slice/crio-498f44e558fbd34d8865d29469eb430458fbcd5e3192f9caad9e81c412b6775e WatchSource:0}: Error finding container 498f44e558fbd34d8865d29469eb430458fbcd5e3192f9caad9e81c412b6775e: Status 404 returned error can't find the container with id 498f44e558fbd34d8865d29469eb430458fbcd5e3192f9caad9e81c412b6775e Mar 18 10:25:36 crc kubenswrapper[4733]: I0318 10:25:36.684098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-rd2dh" event={"ID":"ce77d29d-b82e-46be-a694-b6eea5da9379","Type":"ContainerStarted","Data":"670eec05213c29a80328173b18779c9fe6a9cb2e660ea26278fbeec449be53fe"} Mar 18 10:25:36 crc kubenswrapper[4733]: I0318 10:25:36.686929 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" event={"ID":"585c06be-86bd-48b7-954e-9aec01b08874","Type":"ContainerStarted","Data":"5c8591c08d39a2c1dcf39bc2b94cf19e43b4ae811739fc182daf63959295a448"} Mar 18 10:25:36 crc kubenswrapper[4733]: I0318 10:25:36.689668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" event={"ID":"534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8","Type":"ContainerStarted","Data":"498f44e558fbd34d8865d29469eb430458fbcd5e3192f9caad9e81c412b6775e"} Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.720332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" event={"ID":"585c06be-86bd-48b7-954e-9aec01b08874","Type":"ContainerStarted","Data":"08df0d2f67ec74d809ebd16ed32107787b740a0e4aeff6dcfb9418a3464f8be9"} Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.722422 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" event={"ID":"534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8","Type":"ContainerStarted","Data":"adb53a592c227c2d4b28596b60bab28de1a00c8006702f74cfbf7e0127b715b9"} Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.722559 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.724451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-rd2dh" event={"ID":"ce77d29d-b82e-46be-a694-b6eea5da9379","Type":"ContainerStarted","Data":"d9fed630241159e6db14d549fb1efe33161d7357f061ec3aa85a6de0b53aa9a0"} Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.743085 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-8ds68" podStartSLOduration=1.8752565460000001 podStartE2EDuration="5.743071781s" podCreationTimestamp="2026-03-18 10:25:35 +0000 UTC" firstStartedPulling="2026-03-18 10:25:36.021523854 +0000 UTC m=+775.513258179" lastFinishedPulling="2026-03-18 10:25:39.889339079 +0000 UTC m=+779.381073414" observedRunningTime="2026-03-18 10:25:40.740704094 +0000 UTC m=+780.232438429" watchObservedRunningTime="2026-03-18 10:25:40.743071781 +0000 UTC m=+780.234806106" Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.763678 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" podStartSLOduration=1.989877022 podStartE2EDuration="5.76366297s" podCreationTimestamp="2026-03-18 10:25:35 +0000 UTC" firstStartedPulling="2026-03-18 10:25:36.034011238 +0000 UTC m=+775.525745563" lastFinishedPulling="2026-03-18 10:25:39.807797146 +0000 UTC m=+779.299531511" observedRunningTime="2026-03-18 10:25:40.763413753 +0000 UTC m=+780.255148088" watchObservedRunningTime="2026-03-18 10:25:40.76366297 +0000 UTC m=+780.255397285" Mar 18 10:25:40 crc kubenswrapper[4733]: I0318 10:25:40.794111 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-rd2dh" podStartSLOduration=1.756700613 podStartE2EDuration="5.794081959s" podCreationTimestamp="2026-03-18 10:25:35 +0000 UTC" firstStartedPulling="2026-03-18 10:25:35.769613747 +0000 UTC m=+775.261348072" lastFinishedPulling="2026-03-18 10:25:39.806995093 +0000 UTC m=+779.298729418" observedRunningTime="2026-03-18 10:25:40.787424088 +0000 UTC m=+780.279158413" watchObservedRunningTime="2026-03-18 10:25:40.794081959 +0000 UTC m=+780.285816324" Mar 18 10:25:45 crc kubenswrapper[4733]: I0318 10:25:45.504586 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-2nr27" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.351631 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pxwd"] Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353261 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="nbdb" containerID="cri-o://10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353403 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-acl-logging" containerID="cri-o://9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353293 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353497 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-node" containerID="cri-o://8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353227 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-controller" containerID="cri-o://e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353579 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="sbdb" containerID="cri-o://de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.353324 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="northd" containerID="cri-o://3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.412804 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" containerID="cri-o://850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" gracePeriod=30 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.734396 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/3.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.742683 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovn-acl-logging/0.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.743688 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovn-controller/0.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.744736 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783751 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783798 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783831 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783907 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783944 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.783971 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784000 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784075 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784108 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784160 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784229 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784282 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784318 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784350 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784379 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.784411 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqxdr\" (UniqueName: \"kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr\") pod \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\" (UID: \"73327417-4d3b-45f1-b3b6-575fdeeaa31a\") " Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785050 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket" (OuterVolumeSpecName: "log-socket") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785228 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785301 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785343 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785755 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785812 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785851 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785887 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log" (OuterVolumeSpecName: "node-log") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785873 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785914 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.785970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.786008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.786047 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash" (OuterVolumeSpecName: "host-slash") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.786631 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.786896 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.787468 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.794310 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr" (OuterVolumeSpecName: "kube-api-access-zqxdr") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "kube-api-access-zqxdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.796830 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.805268 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "73327417-4d3b-45f1-b3b6-575fdeeaa31a" (UID: "73327417-4d3b-45f1-b3b6-575fdeeaa31a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.815316 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/2.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.815965 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/1.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.816034 4733 generic.go:334] "Generic (PLEG): container finished" podID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" containerID="e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d" exitCode=2 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.816129 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerDied","Data":"e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.816182 4733 scope.go:117] "RemoveContainer" containerID="b6a4e9643a717b3f38fc1bed5c534e12bb873f0ffcf3c504cb4395c11621a73a" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.817370 4733 scope.go:117] "RemoveContainer" containerID="e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.818830 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-g6j2q_openshift-multus(cc85b0d4-15a5-4894-9f07-9aaeb28f63fa)\"" pod="openshift-multus/multus-g6j2q" podUID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.823449 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovnkube-controller/3.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.827122 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovn-acl-logging/0.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.828015 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7pxwd_73327417-4d3b-45f1-b3b6-575fdeeaa31a/ovn-controller/0.log" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.829565 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.829803 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.829978 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.830132 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.830328 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.830476 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" exitCode=0 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.830667 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" exitCode=143 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.830854 4733 generic.go:334] "Generic (PLEG): container finished" podID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" exitCode=143 Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.829717 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831145 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831315 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831587 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.831855 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832000 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832133 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832289 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832412 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832523 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832620 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832749 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832872 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.832987 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833103 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833262 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833406 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833522 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833620 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833728 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833854 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.833969 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.834077 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.834220 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.834344 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.834463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.834763 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.835033 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.835451 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.835670 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.835881 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.836093 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.836333 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.836532 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.836733 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.836997 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.837164 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" event={"ID":"73327417-4d3b-45f1-b3b6-575fdeeaa31a","Type":"ContainerDied","Data":"35bea9a3e63456f3c4522f7b18c54f2df3fc823d29bd3059264ea8e5f121d012"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.837333 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.837452 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.837592 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.837755 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838107 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838258 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838457 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838700 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838874 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.838997 4733 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.839121 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pwtdp"] Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.829695 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7pxwd" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.839795 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.839937 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.840056 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.840341 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.841015 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.841160 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.841321 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="northd" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.841427 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="northd" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.841551 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.841676 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.841858 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="nbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.841980 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="nbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.842095 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="sbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.842236 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="sbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.842376 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-acl-logging" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.842492 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-acl-logging" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.842836 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-node" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.842974 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-node" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.843096 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kubecfg-setup" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.843239 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kubecfg-setup" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.843374 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.843489 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.843810 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="northd" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.843965 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844083 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844254 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844371 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="kube-rbac-proxy-node" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844503 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844611 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="sbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844729 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-acl-logging" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.844856 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.845143 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovn-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.845339 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.845451 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="nbdb" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.845843 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.845996 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: E0318 10:25:51.847078 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.847308 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" containerName="ovnkube-controller" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.850117 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.851965 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.854633 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.855407 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.865756 4733 scope.go:117] "RemoveContainer" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.887912 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-bin\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.888032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-env-overrides\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.888068 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-kubelet\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.888102 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24b8k\" (UniqueName: \"kubernetes.io/projected/a10a6717-677d-43b8-8d78-8f60b16836ed-kube-api-access-24b8k\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.888887 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-ovn\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.889000 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-config\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.889067 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-systemd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.889098 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-script-lib\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.889137 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-netd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.889168 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-etc-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891334 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-node-log\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891673 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-var-lib-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891745 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891824 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-systemd-units\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891878 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.891956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a10a6717-677d-43b8-8d78-8f60b16836ed-ovn-node-metrics-cert\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892077 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-netns\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892141 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-slash\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-log-socket\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892645 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892750 4733 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-node-log\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892824 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892900 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.892993 4733 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893094 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893172 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893278 4733 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893359 4733 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893435 4733 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-slash\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893515 4733 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893595 4733 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893736 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893848 4733 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.893943 4733 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-log-socket\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.894038 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.894152 4733 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.894277 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/73327417-4d3b-45f1-b3b6-575fdeeaa31a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.894396 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/73327417-4d3b-45f1-b3b6-575fdeeaa31a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.894573 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqxdr\" (UniqueName: \"kubernetes.io/projected/73327417-4d3b-45f1-b3b6-575fdeeaa31a-kube-api-access-zqxdr\") on node \"crc\" DevicePath \"\"" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.897375 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.914927 4733 scope.go:117] "RemoveContainer" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.932009 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pxwd"] Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.937692 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7pxwd"] Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.944311 4733 scope.go:117] "RemoveContainer" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.956879 4733 scope.go:117] "RemoveContainer" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.967819 4733 scope.go:117] "RemoveContainer" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.987437 4733 scope.go:117] "RemoveContainer" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995800 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a10a6717-677d-43b8-8d78-8f60b16836ed-ovn-node-metrics-cert\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995840 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-netns\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995852 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995870 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-slash\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-log-socket\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-bin\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-env-overrides\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996393 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24b8k\" (UniqueName: \"kubernetes.io/projected/a10a6717-677d-43b8-8d78-8f60b16836ed-kube-api-access-24b8k\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-kubelet\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996457 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-ovn\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996519 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-config\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996563 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-systemd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996596 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-script-lib\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996634 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-netd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-node-log\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996694 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-etc-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-var-lib-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996811 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-systemd-units\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996908 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-systemd-units\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.996970 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-netd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-node-log\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997035 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-env-overrides\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997061 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-etc-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995910 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-slash\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997090 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-log-socket\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995932 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997123 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-var-lib-openvswitch\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997165 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-ovn\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.995983 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-netns\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997254 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-run-ovn-kubernetes\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997130 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-cni-bin\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997417 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-host-kubelet\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997448 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a10a6717-677d-43b8-8d78-8f60b16836ed-run-systemd\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.997865 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-script-lib\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.998401 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a10a6717-677d-43b8-8d78-8f60b16836ed-ovnkube-config\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.998994 4733 scope.go:117] "RemoveContainer" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:51 crc kubenswrapper[4733]: I0318 10:25:51.999345 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a10a6717-677d-43b8-8d78-8f60b16836ed-ovn-node-metrics-cert\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.013507 4733 scope.go:117] "RemoveContainer" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.016814 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24b8k\" (UniqueName: \"kubernetes.io/projected/a10a6717-677d-43b8-8d78-8f60b16836ed-kube-api-access-24b8k\") pod \"ovnkube-node-pwtdp\" (UID: \"a10a6717-677d-43b8-8d78-8f60b16836ed\") " pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.032783 4733 scope.go:117] "RemoveContainer" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.053403 4733 scope.go:117] "RemoveContainer" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.054035 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": container with ID starting with 850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395 not found: ID does not exist" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.054078 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} err="failed to get container status \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": rpc error: code = NotFound desc = could not find container \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": container with ID starting with 850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.054111 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.054713 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": container with ID starting with f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291 not found: ID does not exist" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.054750 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} err="failed to get container status \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": rpc error: code = NotFound desc = could not find container \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": container with ID starting with f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.054776 4733 scope.go:117] "RemoveContainer" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.055427 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": container with ID starting with de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291 not found: ID does not exist" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.055463 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} err="failed to get container status \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": rpc error: code = NotFound desc = could not find container \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": container with ID starting with de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.055490 4733 scope.go:117] "RemoveContainer" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.057355 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": container with ID starting with 10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77 not found: ID does not exist" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.057391 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} err="failed to get container status \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": rpc error: code = NotFound desc = could not find container \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": container with ID starting with 10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.057422 4733 scope.go:117] "RemoveContainer" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.058065 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": container with ID starting with 3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea not found: ID does not exist" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.058101 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} err="failed to get container status \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": rpc error: code = NotFound desc = could not find container \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": container with ID starting with 3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.058125 4733 scope.go:117] "RemoveContainer" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.058634 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": container with ID starting with c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1 not found: ID does not exist" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.058665 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} err="failed to get container status \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": rpc error: code = NotFound desc = could not find container \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": container with ID starting with c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.058687 4733 scope.go:117] "RemoveContainer" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.058985 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": container with ID starting with 8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4 not found: ID does not exist" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059008 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} err="failed to get container status \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": rpc error: code = NotFound desc = could not find container \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": container with ID starting with 8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059025 4733 scope.go:117] "RemoveContainer" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.059268 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": container with ID starting with 9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226 not found: ID does not exist" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059308 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} err="failed to get container status \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": rpc error: code = NotFound desc = could not find container \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": container with ID starting with 9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059330 4733 scope.go:117] "RemoveContainer" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.059573 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": container with ID starting with e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503 not found: ID does not exist" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059612 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} err="failed to get container status \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": rpc error: code = NotFound desc = could not find container \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": container with ID starting with e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059642 4733 scope.go:117] "RemoveContainer" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: E0318 10:25:52.059917 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": container with ID starting with d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378 not found: ID does not exist" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059947 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} err="failed to get container status \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": rpc error: code = NotFound desc = could not find container \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": container with ID starting with d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.059970 4733 scope.go:117] "RemoveContainer" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.060433 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} err="failed to get container status \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": rpc error: code = NotFound desc = could not find container \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": container with ID starting with 850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.060470 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061002 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} err="failed to get container status \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": rpc error: code = NotFound desc = could not find container \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": container with ID starting with f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061053 4733 scope.go:117] "RemoveContainer" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061405 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} err="failed to get container status \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": rpc error: code = NotFound desc = could not find container \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": container with ID starting with de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061440 4733 scope.go:117] "RemoveContainer" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061888 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} err="failed to get container status \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": rpc error: code = NotFound desc = could not find container \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": container with ID starting with 10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.061929 4733 scope.go:117] "RemoveContainer" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.062763 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} err="failed to get container status \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": rpc error: code = NotFound desc = could not find container \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": container with ID starting with 3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.062799 4733 scope.go:117] "RemoveContainer" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.063361 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} err="failed to get container status \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": rpc error: code = NotFound desc = could not find container \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": container with ID starting with c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.063399 4733 scope.go:117] "RemoveContainer" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.063693 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} err="failed to get container status \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": rpc error: code = NotFound desc = could not find container \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": container with ID starting with 8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.063724 4733 scope.go:117] "RemoveContainer" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064007 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} err="failed to get container status \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": rpc error: code = NotFound desc = could not find container \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": container with ID starting with 9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064033 4733 scope.go:117] "RemoveContainer" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064301 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} err="failed to get container status \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": rpc error: code = NotFound desc = could not find container \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": container with ID starting with e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064326 4733 scope.go:117] "RemoveContainer" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064547 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} err="failed to get container status \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": rpc error: code = NotFound desc = could not find container \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": container with ID starting with d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064570 4733 scope.go:117] "RemoveContainer" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064780 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} err="failed to get container status \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": rpc error: code = NotFound desc = could not find container \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": container with ID starting with 850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.064802 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065008 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} err="failed to get container status \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": rpc error: code = NotFound desc = could not find container \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": container with ID starting with f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065031 4733 scope.go:117] "RemoveContainer" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065274 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} err="failed to get container status \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": rpc error: code = NotFound desc = could not find container \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": container with ID starting with de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065300 4733 scope.go:117] "RemoveContainer" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065516 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} err="failed to get container status \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": rpc error: code = NotFound desc = could not find container \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": container with ID starting with 10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065539 4733 scope.go:117] "RemoveContainer" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065740 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} err="failed to get container status \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": rpc error: code = NotFound desc = could not find container \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": container with ID starting with 3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065763 4733 scope.go:117] "RemoveContainer" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065971 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} err="failed to get container status \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": rpc error: code = NotFound desc = could not find container \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": container with ID starting with c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.065993 4733 scope.go:117] "RemoveContainer" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066253 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} err="failed to get container status \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": rpc error: code = NotFound desc = could not find container \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": container with ID starting with 8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066278 4733 scope.go:117] "RemoveContainer" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066489 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} err="failed to get container status \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": rpc error: code = NotFound desc = could not find container \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": container with ID starting with 9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066512 4733 scope.go:117] "RemoveContainer" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066720 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} err="failed to get container status \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": rpc error: code = NotFound desc = could not find container \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": container with ID starting with e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066743 4733 scope.go:117] "RemoveContainer" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066957 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} err="failed to get container status \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": rpc error: code = NotFound desc = could not find container \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": container with ID starting with d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.066987 4733 scope.go:117] "RemoveContainer" containerID="850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067246 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395"} err="failed to get container status \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": rpc error: code = NotFound desc = could not find container \"850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395\": container with ID starting with 850880b1c00b2f5a5a32f08989e49cc1406960901b41de4ee69b92f38458d395 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067272 4733 scope.go:117] "RemoveContainer" containerID="f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067490 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291"} err="failed to get container status \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": rpc error: code = NotFound desc = could not find container \"f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291\": container with ID starting with f271860bb80800ec82f217effead5b1e9475829bbf78baea857aa7639eea7291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067516 4733 scope.go:117] "RemoveContainer" containerID="de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067724 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291"} err="failed to get container status \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": rpc error: code = NotFound desc = could not find container \"de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291\": container with ID starting with de2d3be12ab406039374efc6a0094e21103be62b51ef65c4ccf5529d6ef05291 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067746 4733 scope.go:117] "RemoveContainer" containerID="10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067962 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77"} err="failed to get container status \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": rpc error: code = NotFound desc = could not find container \"10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77\": container with ID starting with 10bb15e1466307914440acf630eb4ea4a442cf4354a9cc5e6ddb40d8147a4d77 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.067992 4733 scope.go:117] "RemoveContainer" containerID="3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068260 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea"} err="failed to get container status \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": rpc error: code = NotFound desc = could not find container \"3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea\": container with ID starting with 3c7ac487e3e86e35e8a1d6ddf67975eb4a67657b219938a69a90ccd5774ee0ea not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068294 4733 scope.go:117] "RemoveContainer" containerID="c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068535 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1"} err="failed to get container status \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": rpc error: code = NotFound desc = could not find container \"c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1\": container with ID starting with c6233b06ff88fa77ac74becea9ef44fd4aa09b0ae718390c1b73c78d353ecbc1 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068566 4733 scope.go:117] "RemoveContainer" containerID="8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068803 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4"} err="failed to get container status \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": rpc error: code = NotFound desc = could not find container \"8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4\": container with ID starting with 8f1102a1e204850c8494267640da7ec93ec67e7341c4eb60d22a7f0772058cb4 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.068835 4733 scope.go:117] "RemoveContainer" containerID="9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.069065 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226"} err="failed to get container status \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": rpc error: code = NotFound desc = could not find container \"9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226\": container with ID starting with 9eb388a698a6a2ee1963deeba09459d7190f60ed189ee20a2ba24de317604226 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.069091 4733 scope.go:117] "RemoveContainer" containerID="e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.069356 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503"} err="failed to get container status \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": rpc error: code = NotFound desc = could not find container \"e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503\": container with ID starting with e3043711b80807f025d3bd2e7b4593f22d78dd3e458aa185c18c065af4aca503 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.069389 4733 scope.go:117] "RemoveContainer" containerID="d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.069635 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378"} err="failed to get container status \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": rpc error: code = NotFound desc = could not find container \"d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378\": container with ID starting with d8e8b34a9866e756d4bedfedebd589abe90e519eb174e5962e4744d0d6c7f378 not found: ID does not exist" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.180478 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.842895 4733 generic.go:334] "Generic (PLEG): container finished" podID="a10a6717-677d-43b8-8d78-8f60b16836ed" containerID="d4da2b3b60b9de1cafe978e281f7d1a58bcc91654d6d64b25f5b83fe0d977358" exitCode=0 Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.842965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerDied","Data":"d4da2b3b60b9de1cafe978e281f7d1a58bcc91654d6d64b25f5b83fe0d977358"} Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.842988 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"61932908893bc650e446c63cff073523850debe73375b309dc80abfe37eae512"} Mar 18 10:25:52 crc kubenswrapper[4733]: I0318 10:25:52.848079 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/2.log" Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.185285 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73327417-4d3b-45f1-b3b6-575fdeeaa31a" path="/var/lib/kubelet/pods/73327417-4d3b-45f1-b3b6-575fdeeaa31a/volumes" Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.862335 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"a2d893b5c4cf7c22fc1504a2a49212da6cfea5758ab2bc466cac6aa3003a9f92"} Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.862908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"9ac071de97432e33422df8a380f9305452b3361e02459419299a65bb49a7a681"} Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.862942 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"e45684d426968124d1ff7372ac51e178cd6e9a5fdf87a6478b6776c86c417590"} Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.862970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"c4f5d37616b6ca27fb9b1ff3b37f035d03db2a230aa08159945f7c70446a286b"} Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.862995 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"baaee666b1ce55cbef60b96672d5f702aa3e96811a7b8870393b299aa946d496"} Mar 18 10:25:53 crc kubenswrapper[4733]: I0318 10:25:53.863019 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"cea74c0bbf0a34edcc4406c026e508f89cb98477944ee37f0a516e38b1ac1083"} Mar 18 10:25:56 crc kubenswrapper[4733]: I0318 10:25:56.902113 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"7a324d5a2a6e0aa931bdb486b9d81da9aee89d6483988b88c47c792938b3c49c"} Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.928436 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" event={"ID":"a10a6717-677d-43b8-8d78-8f60b16836ed","Type":"ContainerStarted","Data":"2474aca6cf95defd9534c2913ac781f9cbd560505ecc63aec421423f42d9bb6b"} Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.930394 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.930421 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.930431 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.959080 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" podStartSLOduration=7.959060382 podStartE2EDuration="7.959060382s" podCreationTimestamp="2026-03-18 10:25:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:25:58.958567808 +0000 UTC m=+798.450302143" watchObservedRunningTime="2026-03-18 10:25:58.959060382 +0000 UTC m=+798.450794707" Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.964205 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:25:58 crc kubenswrapper[4733]: I0318 10:25:58.966287 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.150624 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563826-tfzqx"] Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.151741 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.156241 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.156660 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.156783 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.167635 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563826-tfzqx"] Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.237022 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp6vj\" (UniqueName: \"kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj\") pod \"auto-csr-approver-29563826-tfzqx\" (UID: \"eb9f28a6-6f4c-440b-abfc-cca26041cbef\") " pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.338732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp6vj\" (UniqueName: \"kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj\") pod \"auto-csr-approver-29563826-tfzqx\" (UID: \"eb9f28a6-6f4c-440b-abfc-cca26041cbef\") " pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.365909 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp6vj\" (UniqueName: \"kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj\") pod \"auto-csr-approver-29563826-tfzqx\" (UID: \"eb9f28a6-6f4c-440b-abfc-cca26041cbef\") " pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.477798 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.520820 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(cb18a9006cb61d0feef0c6ba70e5ea60694d961bc6d25656f016d8e25a763ae7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.521382 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(cb18a9006cb61d0feef0c6ba70e5ea60694d961bc6d25656f016d8e25a763ae7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.521426 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(cb18a9006cb61d0feef0c6ba70e5ea60694d961bc6d25656f016d8e25a763ae7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.521554 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(cb18a9006cb61d0feef0c6ba70e5ea60694d961bc6d25656f016d8e25a763ae7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.941128 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: I0318 10:26:00.941602 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.975239 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(aa2611edfeef07d4211957559e0fe8996517afcd8ee9dd4dca6e796dc9e459fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.975309 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(aa2611edfeef07d4211957559e0fe8996517afcd8ee9dd4dca6e796dc9e459fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.975335 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(aa2611edfeef07d4211957559e0fe8996517afcd8ee9dd4dca6e796dc9e459fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:00 crc kubenswrapper[4733]: E0318 10:26:00.975402 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(aa2611edfeef07d4211957559e0fe8996517afcd8ee9dd4dca6e796dc9e459fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" Mar 18 10:26:05 crc kubenswrapper[4733]: I0318 10:26:05.175418 4733 scope.go:117] "RemoveContainer" containerID="e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d" Mar 18 10:26:05 crc kubenswrapper[4733]: E0318 10:26:05.176712 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-g6j2q_openshift-multus(cc85b0d4-15a5-4894-9f07-9aaeb28f63fa)\"" pod="openshift-multus/multus-g6j2q" podUID="cc85b0d4-15a5-4894-9f07-9aaeb28f63fa" Mar 18 10:26:13 crc kubenswrapper[4733]: I0318 10:26:13.571727 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:26:13 crc kubenswrapper[4733]: I0318 10:26:13.572820 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:26:14 crc kubenswrapper[4733]: I0318 10:26:14.174901 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:14 crc kubenswrapper[4733]: I0318 10:26:14.175723 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:14 crc kubenswrapper[4733]: E0318 10:26:14.216279 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(fe870290ef5184b8c09b344f1f32f093def53a2f6d05b152aa0dbc0f6af9f396): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 10:26:14 crc kubenswrapper[4733]: E0318 10:26:14.216389 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(fe870290ef5184b8c09b344f1f32f093def53a2f6d05b152aa0dbc0f6af9f396): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:14 crc kubenswrapper[4733]: E0318 10:26:14.216435 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(fe870290ef5184b8c09b344f1f32f093def53a2f6d05b152aa0dbc0f6af9f396): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:14 crc kubenswrapper[4733]: E0318 10:26:14.216527 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29563826-tfzqx_openshift-infra(eb9f28a6-6f4c-440b-abfc-cca26041cbef)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29563826-tfzqx_openshift-infra_eb9f28a6-6f4c-440b-abfc-cca26041cbef_0(fe870290ef5184b8c09b344f1f32f093def53a2f6d05b152aa0dbc0f6af9f396): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" Mar 18 10:26:18 crc kubenswrapper[4733]: I0318 10:26:18.176257 4733 scope.go:117] "RemoveContainer" containerID="e6e4d066d930397d09ab341b832e9b1659ca8d82f0e6fdc83f2d3f3738f5c64d" Mar 18 10:26:19 crc kubenswrapper[4733]: I0318 10:26:19.078902 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g6j2q_cc85b0d4-15a5-4894-9f07-9aaeb28f63fa/kube-multus/2.log" Mar 18 10:26:19 crc kubenswrapper[4733]: I0318 10:26:19.079350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g6j2q" event={"ID":"cc85b0d4-15a5-4894-9f07-9aaeb28f63fa","Type":"ContainerStarted","Data":"2634f8ab38a5754851b399ebb93dc944c97649d7f30f22a1e2664690641f0fa7"} Mar 18 10:26:22 crc kubenswrapper[4733]: I0318 10:26:22.216925 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pwtdp" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.039034 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p"] Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.040374 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.042966 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.050584 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p"] Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.117906 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.117989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlzb5\" (UniqueName: \"kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.118041 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.219320 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.219389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlzb5\" (UniqueName: \"kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.219434 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.221550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.221961 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.256805 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlzb5\" (UniqueName: \"kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.360658 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:23 crc kubenswrapper[4733]: I0318 10:26:23.681337 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p"] Mar 18 10:26:24 crc kubenswrapper[4733]: I0318 10:26:24.125400 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerStarted","Data":"6c79fe35f870da8b106890820b1a9cf7106b40a2dceff5de739a0c9396b65f31"} Mar 18 10:26:24 crc kubenswrapper[4733]: I0318 10:26:24.125896 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerStarted","Data":"115707dd95d204d5c85fc33da24d6f18d091d0e56db5ebe637e2d7aaa0ea6ad1"} Mar 18 10:26:26 crc kubenswrapper[4733]: I0318 10:26:26.141093 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerID="6c79fe35f870da8b106890820b1a9cf7106b40a2dceff5de739a0c9396b65f31" exitCode=0 Mar 18 10:26:26 crc kubenswrapper[4733]: I0318 10:26:26.141257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerDied","Data":"6c79fe35f870da8b106890820b1a9cf7106b40a2dceff5de739a0c9396b65f31"} Mar 18 10:26:28 crc kubenswrapper[4733]: I0318 10:26:28.159178 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerID="f7502e815678bd10e8b02876d4de520d25734e6bcf12503e9d9d6e0c0c015342" exitCode=0 Mar 18 10:26:28 crc kubenswrapper[4733]: I0318 10:26:28.159714 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerDied","Data":"f7502e815678bd10e8b02876d4de520d25734e6bcf12503e9d9d6e0c0c015342"} Mar 18 10:26:28 crc kubenswrapper[4733]: I0318 10:26:28.175465 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:28 crc kubenswrapper[4733]: I0318 10:26:28.176420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:28 crc kubenswrapper[4733]: I0318 10:26:28.476915 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563826-tfzqx"] Mar 18 10:26:28 crc kubenswrapper[4733]: W0318 10:26:28.490162 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb9f28a6_6f4c_440b_abfc_cca26041cbef.slice/crio-9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6 WatchSource:0}: Error finding container 9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6: Status 404 returned error can't find the container with id 9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6 Mar 18 10:26:29 crc kubenswrapper[4733]: I0318 10:26:29.172467 4733 generic.go:334] "Generic (PLEG): container finished" podID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerID="aaa730e8d1687e25ca04b79b8603e52bb0f970b31d74876a63a754170920f8b9" exitCode=0 Mar 18 10:26:29 crc kubenswrapper[4733]: I0318 10:26:29.173092 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerDied","Data":"aaa730e8d1687e25ca04b79b8603e52bb0f970b31d74876a63a754170920f8b9"} Mar 18 10:26:29 crc kubenswrapper[4733]: I0318 10:26:29.175077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" event={"ID":"eb9f28a6-6f4c-440b-abfc-cca26041cbef","Type":"ContainerStarted","Data":"9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6"} Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.184355 4733 generic.go:334] "Generic (PLEG): container finished" podID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" containerID="73a17ce4bce512adc8ff4282e561fca0880aa24a1a28aaa52332d077a8673f8c" exitCode=0 Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.184430 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" event={"ID":"eb9f28a6-6f4c-440b-abfc-cca26041cbef","Type":"ContainerDied","Data":"73a17ce4bce512adc8ff4282e561fca0880aa24a1a28aaa52332d077a8673f8c"} Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.495248 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.635486 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util\") pod \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.635581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlzb5\" (UniqueName: \"kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5\") pod \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.635651 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle\") pod \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\" (UID: \"3f95562e-ae03-4b2d-92b7-bc5593785f3c\") " Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.636713 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle" (OuterVolumeSpecName: "bundle") pod "3f95562e-ae03-4b2d-92b7-bc5593785f3c" (UID: "3f95562e-ae03-4b2d-92b7-bc5593785f3c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.643521 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5" (OuterVolumeSpecName: "kube-api-access-qlzb5") pod "3f95562e-ae03-4b2d-92b7-bc5593785f3c" (UID: "3f95562e-ae03-4b2d-92b7-bc5593785f3c"). InnerVolumeSpecName "kube-api-access-qlzb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.654828 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util" (OuterVolumeSpecName: "util") pod "3f95562e-ae03-4b2d-92b7-bc5593785f3c" (UID: "3f95562e-ae03-4b2d-92b7-bc5593785f3c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.737301 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-util\") on node \"crc\" DevicePath \"\"" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.737356 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlzb5\" (UniqueName: \"kubernetes.io/projected/3f95562e-ae03-4b2d-92b7-bc5593785f3c-kube-api-access-qlzb5\") on node \"crc\" DevicePath \"\"" Mar 18 10:26:30 crc kubenswrapper[4733]: I0318 10:26:30.737378 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3f95562e-ae03-4b2d-92b7-bc5593785f3c-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.199021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" event={"ID":"3f95562e-ae03-4b2d-92b7-bc5593785f3c","Type":"ContainerDied","Data":"115707dd95d204d5c85fc33da24d6f18d091d0e56db5ebe637e2d7aaa0ea6ad1"} Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.199552 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="115707dd95d204d5c85fc33da24d6f18d091d0e56db5ebe637e2d7aaa0ea6ad1" Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.199038 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p" Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.517236 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.549581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp6vj\" (UniqueName: \"kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj\") pod \"eb9f28a6-6f4c-440b-abfc-cca26041cbef\" (UID: \"eb9f28a6-6f4c-440b-abfc-cca26041cbef\") " Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.556198 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj" (OuterVolumeSpecName: "kube-api-access-bp6vj") pod "eb9f28a6-6f4c-440b-abfc-cca26041cbef" (UID: "eb9f28a6-6f4c-440b-abfc-cca26041cbef"). InnerVolumeSpecName "kube-api-access-bp6vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:26:31 crc kubenswrapper[4733]: I0318 10:26:31.650960 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp6vj\" (UniqueName: \"kubernetes.io/projected/eb9f28a6-6f4c-440b-abfc-cca26041cbef-kube-api-access-bp6vj\") on node \"crc\" DevicePath \"\"" Mar 18 10:26:32 crc kubenswrapper[4733]: I0318 10:26:32.209548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" event={"ID":"eb9f28a6-6f4c-440b-abfc-cca26041cbef","Type":"ContainerDied","Data":"9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6"} Mar 18 10:26:32 crc kubenswrapper[4733]: I0318 10:26:32.209992 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f8f68c9457c58f5aad92937557aa65f4fad0e1186bd7002a29cc99219d0aec6" Mar 18 10:26:32 crc kubenswrapper[4733]: I0318 10:26:32.210100 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563826-tfzqx" Mar 18 10:26:32 crc kubenswrapper[4733]: I0318 10:26:32.612637 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563820-x8mq8"] Mar 18 10:26:32 crc kubenswrapper[4733]: I0318 10:26:32.619266 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563820-x8mq8"] Mar 18 10:26:33 crc kubenswrapper[4733]: I0318 10:26:33.190302 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949d71ae-f754-4b5c-8c0b-fec8d374f27e" path="/var/lib/kubelet/pods/949d71ae-f754-4b5c-8c0b-fec8d374f27e/volumes" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449003 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp"] Mar 18 10:26:34 crc kubenswrapper[4733]: E0318 10:26:34.449237 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="pull" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449248 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="pull" Mar 18 10:26:34 crc kubenswrapper[4733]: E0318 10:26:34.449255 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" containerName="oc" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449261 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" containerName="oc" Mar 18 10:26:34 crc kubenswrapper[4733]: E0318 10:26:34.449275 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="extract" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449282 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="extract" Mar 18 10:26:34 crc kubenswrapper[4733]: E0318 10:26:34.449291 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="util" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449297 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="util" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449392 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" containerName="oc" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449402 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f95562e-ae03-4b2d-92b7-bc5593785f3c" containerName="extract" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.449779 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.452258 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.452414 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.452808 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-kdvvs" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.461299 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp"] Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.519315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbp2v\" (UniqueName: \"kubernetes.io/projected/7c8f098b-42c0-4132-88c0-350e0c872f9d-kube-api-access-wbp2v\") pod \"nmstate-operator-796d4cfff4-2d4dp\" (UID: \"7c8f098b-42c0-4132-88c0-350e0c872f9d\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.620325 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbp2v\" (UniqueName: \"kubernetes.io/projected/7c8f098b-42c0-4132-88c0-350e0c872f9d-kube-api-access-wbp2v\") pod \"nmstate-operator-796d4cfff4-2d4dp\" (UID: \"7c8f098b-42c0-4132-88c0-350e0c872f9d\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.638786 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbp2v\" (UniqueName: \"kubernetes.io/projected/7c8f098b-42c0-4132-88c0-350e0c872f9d-kube-api-access-wbp2v\") pod \"nmstate-operator-796d4cfff4-2d4dp\" (UID: \"7c8f098b-42c0-4132-88c0-350e0c872f9d\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" Mar 18 10:26:34 crc kubenswrapper[4733]: I0318 10:26:34.824404 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" Mar 18 10:26:35 crc kubenswrapper[4733]: I0318 10:26:35.068254 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp"] Mar 18 10:26:35 crc kubenswrapper[4733]: I0318 10:26:35.250466 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" event={"ID":"7c8f098b-42c0-4132-88c0-350e0c872f9d","Type":"ContainerStarted","Data":"55b8034339e1f47af10841a0fb6d53aed4dbdcf3f614f4d6e30878c8f218ef41"} Mar 18 10:26:38 crc kubenswrapper[4733]: I0318 10:26:38.277857 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" event={"ID":"7c8f098b-42c0-4132-88c0-350e0c872f9d","Type":"ContainerStarted","Data":"d7d6ef59a7ccc94cd571a4f6c8fa734e4d332d1da6eabe513015c2090c31f499"} Mar 18 10:26:38 crc kubenswrapper[4733]: I0318 10:26:38.300416 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-2d4dp" podStartSLOduration=1.9651879270000001 podStartE2EDuration="4.300387377s" podCreationTimestamp="2026-03-18 10:26:34 +0000 UTC" firstStartedPulling="2026-03-18 10:26:35.08124262 +0000 UTC m=+834.572976955" lastFinishedPulling="2026-03-18 10:26:37.41644208 +0000 UTC m=+836.908176405" observedRunningTime="2026-03-18 10:26:38.294681294 +0000 UTC m=+837.786415689" watchObservedRunningTime="2026-03-18 10:26:38.300387377 +0000 UTC m=+837.792121742" Mar 18 10:26:40 crc kubenswrapper[4733]: I0318 10:26:40.038561 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.000949 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.003261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.005337 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-lcxp9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.043737 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-8jncr"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.046914 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.061592 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.062550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.064473 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.065803 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.080433 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.147396 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.148848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-nmstate-lock\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.148927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg4n7\" (UniqueName: \"kubernetes.io/projected/4c5d76ae-c917-4ba7-91d7-332a8e578245-kube-api-access-lg4n7\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.148955 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-dbus-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.148974 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-ovs-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.149001 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nj9z\" (UniqueName: \"kubernetes.io/projected/eb2e5225-c943-4b06-b2de-90ab1168242b-kube-api-access-5nj9z\") pod \"nmstate-metrics-9b8c8685d-7swn6\" (UID: \"eb2e5225-c943-4b06-b2de-90ab1168242b\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.151364 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.153404 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-l7t6j" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.155635 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.155879 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.161819 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250010 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nj9z\" (UniqueName: \"kubernetes.io/projected/eb2e5225-c943-4b06-b2de-90ab1168242b-kube-api-access-5nj9z\") pod \"nmstate-metrics-9b8c8685d-7swn6\" (UID: \"eb2e5225-c943-4b06-b2de-90ab1168242b\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-nmstate-lock\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250469 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29460af7-7801-4268-aae8-f84763762e2f-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250500 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5fkq\" (UniqueName: \"kubernetes.io/projected/29460af7-7801-4268-aae8-f84763762e2f-kube-api-access-p5fkq\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg4n7\" (UniqueName: \"kubernetes.io/projected/4c5d76ae-c917-4ba7-91d7-332a8e578245-kube-api-access-lg4n7\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250626 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-nmstate-lock\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250816 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-dbus-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250852 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-ovs-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.250927 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-ovs-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.251136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4c5d76ae-c917-4ba7-91d7-332a8e578245-dbus-socket\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.273625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nj9z\" (UniqueName: \"kubernetes.io/projected/eb2e5225-c943-4b06-b2de-90ab1168242b-kube-api-access-5nj9z\") pod \"nmstate-metrics-9b8c8685d-7swn6\" (UID: \"eb2e5225-c943-4b06-b2de-90ab1168242b\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.273753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg4n7\" (UniqueName: \"kubernetes.io/projected/4c5d76ae-c917-4ba7-91d7-332a8e578245-kube-api-access-lg4n7\") pod \"nmstate-handler-8jncr\" (UID: \"4c5d76ae-c917-4ba7-91d7-332a8e578245\") " pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.343985 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5dd4b96b5d-zqmlh"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.344735 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.347667 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351744 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-oauth-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-oauth-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351849 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qlmx\" (UniqueName: \"kubernetes.io/projected/95b678ac-c7be-4c57-8663-05b207f43338-kube-api-access-5qlmx\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351886 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29460af7-7801-4268-aae8-f84763762e2f-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351910 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5fkq\" (UniqueName: \"kubernetes.io/projected/29460af7-7801-4268-aae8-f84763762e2f-kube-api-access-p5fkq\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351930 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-service-ca\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.351971 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.352003 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-console-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.352023 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm52g\" (UniqueName: \"kubernetes.io/projected/945cd091-c09f-4763-8093-ba83e642949f-kube-api-access-sm52g\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.352043 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/95b678ac-c7be-4c57-8663-05b207f43338-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.352062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-trusted-ca-bundle\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.352087 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/95b678ac-c7be-4c57-8663-05b207f43338-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.357947 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29460af7-7801-4268-aae8-f84763762e2f-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.358684 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dd4b96b5d-zqmlh"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.370650 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5fkq\" (UniqueName: \"kubernetes.io/projected/29460af7-7801-4268-aae8-f84763762e2f-kube-api-access-p5fkq\") pod \"nmstate-webhook-5f558f5558-m6rhx\" (UID: \"29460af7-7801-4268-aae8-f84763762e2f\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.373225 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.378474 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453373 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-oauth-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453724 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-oauth-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qlmx\" (UniqueName: \"kubernetes.io/projected/95b678ac-c7be-4c57-8663-05b207f43338-kube-api-access-5qlmx\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453795 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-service-ca\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453859 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-console-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453873 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm52g\" (UniqueName: \"kubernetes.io/projected/945cd091-c09f-4763-8093-ba83e642949f-kube-api-access-sm52g\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/95b678ac-c7be-4c57-8663-05b207f43338-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-trusted-ca-bundle\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.453945 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/95b678ac-c7be-4c57-8663-05b207f43338-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.454541 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-oauth-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.454869 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/95b678ac-c7be-4c57-8663-05b207f43338-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.456072 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-console-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.457318 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-service-ca\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.457325 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/945cd091-c09f-4763-8093-ba83e642949f-trusted-ca-bundle\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.462368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/95b678ac-c7be-4c57-8663-05b207f43338-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.470475 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-serving-cert\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.475463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qlmx\" (UniqueName: \"kubernetes.io/projected/95b678ac-c7be-4c57-8663-05b207f43338-kube-api-access-5qlmx\") pod \"nmstate-console-plugin-86f58fcf4-btpf9\" (UID: \"95b678ac-c7be-4c57-8663-05b207f43338\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.475720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/945cd091-c09f-4763-8093-ba83e642949f-console-oauth-config\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.477060 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm52g\" (UniqueName: \"kubernetes.io/projected/945cd091-c09f-4763-8093-ba83e642949f-kube-api-access-sm52g\") pod \"console-5dd4b96b5d-zqmlh\" (UID: \"945cd091-c09f-4763-8093-ba83e642949f\") " pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.570839 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.570893 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.707223 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.771940 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.802939 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.845432 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx"] Mar 18 10:26:43 crc kubenswrapper[4733]: I0318 10:26:43.949104 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dd4b96b5d-zqmlh"] Mar 18 10:26:43 crc kubenswrapper[4733]: W0318 10:26:43.955388 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod945cd091_c09f_4763_8093_ba83e642949f.slice/crio-0fcbc5b2715fb945771bb4e45ca8e97ef927621c516ac4545d113d30b21dbe45 WatchSource:0}: Error finding container 0fcbc5b2715fb945771bb4e45ca8e97ef927621c516ac4545d113d30b21dbe45: Status 404 returned error can't find the container with id 0fcbc5b2715fb945771bb4e45ca8e97ef927621c516ac4545d113d30b21dbe45 Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.220729 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9"] Mar 18 10:26:44 crc kubenswrapper[4733]: W0318 10:26:44.222769 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95b678ac_c7be_4c57_8663_05b207f43338.slice/crio-1dc84b8c6e3d970faa0bdbd3e8648b36dd5718802988d88510f633589e313fde WatchSource:0}: Error finding container 1dc84b8c6e3d970faa0bdbd3e8648b36dd5718802988d88510f633589e313fde: Status 404 returned error can't find the container with id 1dc84b8c6e3d970faa0bdbd3e8648b36dd5718802988d88510f633589e313fde Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.320691 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8jncr" event={"ID":"4c5d76ae-c917-4ba7-91d7-332a8e578245","Type":"ContainerStarted","Data":"6ab5acedaccba19ca2d61c2a4cb76a0aded93dc197ab1404f37722f281697a54"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.322124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" event={"ID":"eb2e5225-c943-4b06-b2de-90ab1168242b","Type":"ContainerStarted","Data":"5080197f0636affd33fb711f1b386ccebdd88e500e6f1a865cdbbdf72b2bddb2"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.323472 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" event={"ID":"95b678ac-c7be-4c57-8663-05b207f43338","Type":"ContainerStarted","Data":"1dc84b8c6e3d970faa0bdbd3e8648b36dd5718802988d88510f633589e313fde"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.327425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dd4b96b5d-zqmlh" event={"ID":"945cd091-c09f-4763-8093-ba83e642949f","Type":"ContainerStarted","Data":"cc47ca98e8782deb4d798932d2dc218a2467fdeb0abee8feb00454d70f952efa"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.327466 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dd4b96b5d-zqmlh" event={"ID":"945cd091-c09f-4763-8093-ba83e642949f","Type":"ContainerStarted","Data":"0fcbc5b2715fb945771bb4e45ca8e97ef927621c516ac4545d113d30b21dbe45"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.328686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" event={"ID":"29460af7-7801-4268-aae8-f84763762e2f","Type":"ContainerStarted","Data":"c5f05e3d43f450e1955fbac937f68f35f2457a07e96f55f2df71849cb74083ef"} Mar 18 10:26:44 crc kubenswrapper[4733]: I0318 10:26:44.357749 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5dd4b96b5d-zqmlh" podStartSLOduration=1.357724759 podStartE2EDuration="1.357724759s" podCreationTimestamp="2026-03-18 10:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:26:44.353334234 +0000 UTC m=+843.845068569" watchObservedRunningTime="2026-03-18 10:26:44.357724759 +0000 UTC m=+843.849459094" Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.350902 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" event={"ID":"95b678ac-c7be-4c57-8663-05b207f43338","Type":"ContainerStarted","Data":"0812f89430e5b52e2e1cb4addedf266bf709ff4039eff4e0ac0d69576cb60412"} Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.354284 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.354354 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" event={"ID":"29460af7-7801-4268-aae8-f84763762e2f","Type":"ContainerStarted","Data":"ca9eceb62c3c34b30ce951bab0cc3fdbaa6d0e81c4b060a2021b2b230e232828"} Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.358225 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-8jncr" event={"ID":"4c5d76ae-c917-4ba7-91d7-332a8e578245","Type":"ContainerStarted","Data":"e79687014784c6df11db62f5830bad52f1f1bd810a7c455066c5b4c4bee29b5f"} Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.359113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.361228 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" event={"ID":"eb2e5225-c943-4b06-b2de-90ab1168242b","Type":"ContainerStarted","Data":"16abd110e2c3a2397e31e2fbe828abd78cd267f5a803a7d2cdc44df903a8eb78"} Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.380803 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-btpf9" podStartSLOduration=2.03521635 podStartE2EDuration="4.374950288s" podCreationTimestamp="2026-03-18 10:26:43 +0000 UTC" firstStartedPulling="2026-03-18 10:26:44.226153761 +0000 UTC m=+843.717888096" lastFinishedPulling="2026-03-18 10:26:46.565887709 +0000 UTC m=+846.057622034" observedRunningTime="2026-03-18 10:26:47.37185502 +0000 UTC m=+846.863589385" watchObservedRunningTime="2026-03-18 10:26:47.374950288 +0000 UTC m=+846.866684653" Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.397810 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-8jncr" podStartSLOduration=1.349079352 podStartE2EDuration="4.397790601s" podCreationTimestamp="2026-03-18 10:26:43 +0000 UTC" firstStartedPulling="2026-03-18 10:26:43.436885477 +0000 UTC m=+842.928619802" lastFinishedPulling="2026-03-18 10:26:46.485596726 +0000 UTC m=+845.977331051" observedRunningTime="2026-03-18 10:26:47.391776599 +0000 UTC m=+846.883510934" watchObservedRunningTime="2026-03-18 10:26:47.397790601 +0000 UTC m=+846.889524926" Mar 18 10:26:47 crc kubenswrapper[4733]: I0318 10:26:47.413232 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" podStartSLOduration=2.787611978 podStartE2EDuration="5.41317308s" podCreationTimestamp="2026-03-18 10:26:42 +0000 UTC" firstStartedPulling="2026-03-18 10:26:43.862521845 +0000 UTC m=+843.354256160" lastFinishedPulling="2026-03-18 10:26:46.488082937 +0000 UTC m=+845.979817262" observedRunningTime="2026-03-18 10:26:47.410099312 +0000 UTC m=+846.901833657" watchObservedRunningTime="2026-03-18 10:26:47.41317308 +0000 UTC m=+846.904907445" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.431532 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.432981 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.439219 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.628698 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn6vm\" (UniqueName: \"kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.628751 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.628840 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.730468 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.730828 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn6vm\" (UniqueName: \"kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.730855 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.730995 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.731306 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.753678 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn6vm\" (UniqueName: \"kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm\") pod \"community-operators-qpfbg\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:48 crc kubenswrapper[4733]: I0318 10:26:48.781674 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:49 crc kubenswrapper[4733]: I0318 10:26:49.375046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" event={"ID":"eb2e5225-c943-4b06-b2de-90ab1168242b","Type":"ContainerStarted","Data":"42590cc87de02e5705810a3bbc50a741275a61b4ae89fb7d3122b0303f2aea96"} Mar 18 10:26:49 crc kubenswrapper[4733]: I0318 10:26:49.401301 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-7swn6" podStartSLOduration=2.059048798 podStartE2EDuration="7.401270775s" podCreationTimestamp="2026-03-18 10:26:42 +0000 UTC" firstStartedPulling="2026-03-18 10:26:43.818569399 +0000 UTC m=+843.310303744" lastFinishedPulling="2026-03-18 10:26:49.160791386 +0000 UTC m=+848.652525721" observedRunningTime="2026-03-18 10:26:49.38990686 +0000 UTC m=+848.881641225" watchObservedRunningTime="2026-03-18 10:26:49.401270775 +0000 UTC m=+848.893005150" Mar 18 10:26:49 crc kubenswrapper[4733]: I0318 10:26:49.579598 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:26:50 crc kubenswrapper[4733]: I0318 10:26:50.384051 4733 generic.go:334] "Generic (PLEG): container finished" podID="e752ae43-529b-407b-8346-a9eb89990c1f" containerID="c39a8c0e7870cb3625474393180c8aa4297d3c1427acc66e1c577dd7840a186c" exitCode=0 Mar 18 10:26:50 crc kubenswrapper[4733]: I0318 10:26:50.384126 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerDied","Data":"c39a8c0e7870cb3625474393180c8aa4297d3c1427acc66e1c577dd7840a186c"} Mar 18 10:26:50 crc kubenswrapper[4733]: I0318 10:26:50.384695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerStarted","Data":"a969dc374c564078153562a05b2d6e6f30e62624c3adff3d6bf3cbebc8ffe206"} Mar 18 10:26:51 crc kubenswrapper[4733]: I0318 10:26:51.395544 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerStarted","Data":"d84796ff2be656e89a749a5e50910da1185cb7b35004070858a0b58ba51bc5ea"} Mar 18 10:26:52 crc kubenswrapper[4733]: I0318 10:26:52.407867 4733 generic.go:334] "Generic (PLEG): container finished" podID="e752ae43-529b-407b-8346-a9eb89990c1f" containerID="d84796ff2be656e89a749a5e50910da1185cb7b35004070858a0b58ba51bc5ea" exitCode=0 Mar 18 10:26:52 crc kubenswrapper[4733]: I0318 10:26:52.407934 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerDied","Data":"d84796ff2be656e89a749a5e50910da1185cb7b35004070858a0b58ba51bc5ea"} Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.425298 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerStarted","Data":"8a6db4a96078cce01cf2e13028e9a6e04a621dfaaea4705c4e07797befc3b96f"} Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.436093 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-8jncr" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.445841 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.452956 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.454646 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.470492 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qpfbg" podStartSLOduration=3.014091372 podStartE2EDuration="5.470464641s" podCreationTimestamp="2026-03-18 10:26:48 +0000 UTC" firstStartedPulling="2026-03-18 10:26:50.386216138 +0000 UTC m=+849.877950503" lastFinishedPulling="2026-03-18 10:26:52.842589427 +0000 UTC m=+852.334323772" observedRunningTime="2026-03-18 10:26:53.462854354 +0000 UTC m=+852.954588719" watchObservedRunningTime="2026-03-18 10:26:53.470464641 +0000 UTC m=+852.962199006" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.606476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.606543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgw6d\" (UniqueName: \"kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.606649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.707696 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.707927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.708002 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgw6d\" (UniqueName: \"kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.708214 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.708275 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.708333 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.708703 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.716745 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.734557 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgw6d\" (UniqueName: \"kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d\") pod \"redhat-marketplace-5rwz6\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:53 crc kubenswrapper[4733]: I0318 10:26:53.777197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.028526 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.435321 4733 generic.go:334] "Generic (PLEG): container finished" podID="c2eed320-8685-4d5c-8793-e506593a33ca" containerID="e404d64c5fe9561da0e879dd49ffac0a191569c02ab91f0ac47c23b549c8a732" exitCode=0 Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.435382 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerDied","Data":"e404d64c5fe9561da0e879dd49ffac0a191569c02ab91f0ac47c23b549c8a732"} Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.435847 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerStarted","Data":"a61df60801b4e77be52dfd9c59138a93d02124c8a4487a0d508ef12a254c3910"} Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.444058 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5dd4b96b5d-zqmlh" Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.521228 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:26:54 crc kubenswrapper[4733]: I0318 10:26:54.697682 4733 scope.go:117] "RemoveContainer" containerID="4d1f85ec68f66c1e8dcc6134fd20cc9907c6036a83ddad6341fd815f0c10f145" Mar 18 10:26:55 crc kubenswrapper[4733]: I0318 10:26:55.444600 4733 generic.go:334] "Generic (PLEG): container finished" podID="c2eed320-8685-4d5c-8793-e506593a33ca" containerID="c6107d273465cbf0ca33eea0159f3ea649eae43a1b11f06f2154604037aa32c1" exitCode=0 Mar 18 10:26:55 crc kubenswrapper[4733]: I0318 10:26:55.444804 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerDied","Data":"c6107d273465cbf0ca33eea0159f3ea649eae43a1b11f06f2154604037aa32c1"} Mar 18 10:26:56 crc kubenswrapper[4733]: I0318 10:26:56.460294 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerStarted","Data":"c6d8dc1c0b97ce23301426f81a422e6e5a4b20b07fb86d1870204fc64f69c8a8"} Mar 18 10:26:56 crc kubenswrapper[4733]: I0318 10:26:56.491729 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5rwz6" podStartSLOduration=2.074472405 podStartE2EDuration="3.491699415s" podCreationTimestamp="2026-03-18 10:26:53 +0000 UTC" firstStartedPulling="2026-03-18 10:26:54.437199714 +0000 UTC m=+853.928934029" lastFinishedPulling="2026-03-18 10:26:55.854426684 +0000 UTC m=+855.346161039" observedRunningTime="2026-03-18 10:26:56.488086062 +0000 UTC m=+855.979820397" watchObservedRunningTime="2026-03-18 10:26:56.491699415 +0000 UTC m=+855.983433730" Mar 18 10:26:58 crc kubenswrapper[4733]: I0318 10:26:58.785603 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:58 crc kubenswrapper[4733]: I0318 10:26:58.786014 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:58 crc kubenswrapper[4733]: I0318 10:26:58.862871 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:26:59 crc kubenswrapper[4733]: I0318 10:26:59.555720 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.219372 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.220444 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qpfbg" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="registry-server" containerID="cri-o://8a6db4a96078cce01cf2e13028e9a6e04a621dfaaea4705c4e07797befc3b96f" gracePeriod=2 Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.513306 4733 generic.go:334] "Generic (PLEG): container finished" podID="e752ae43-529b-407b-8346-a9eb89990c1f" containerID="8a6db4a96078cce01cf2e13028e9a6e04a621dfaaea4705c4e07797befc3b96f" exitCode=0 Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.513418 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerDied","Data":"8a6db4a96078cce01cf2e13028e9a6e04a621dfaaea4705c4e07797befc3b96f"} Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.674986 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.745123 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities\") pod \"e752ae43-529b-407b-8346-a9eb89990c1f\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.745262 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content\") pod \"e752ae43-529b-407b-8346-a9eb89990c1f\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.745358 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn6vm\" (UniqueName: \"kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm\") pod \"e752ae43-529b-407b-8346-a9eb89990c1f\" (UID: \"e752ae43-529b-407b-8346-a9eb89990c1f\") " Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.746594 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities" (OuterVolumeSpecName: "utilities") pod "e752ae43-529b-407b-8346-a9eb89990c1f" (UID: "e752ae43-529b-407b-8346-a9eb89990c1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.751785 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm" (OuterVolumeSpecName: "kube-api-access-wn6vm") pod "e752ae43-529b-407b-8346-a9eb89990c1f" (UID: "e752ae43-529b-407b-8346-a9eb89990c1f"). InnerVolumeSpecName "kube-api-access-wn6vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.812450 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e752ae43-529b-407b-8346-a9eb89990c1f" (UID: "e752ae43-529b-407b-8346-a9eb89990c1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.847364 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.847407 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn6vm\" (UniqueName: \"kubernetes.io/projected/e752ae43-529b-407b-8346-a9eb89990c1f-kube-api-access-wn6vm\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:02 crc kubenswrapper[4733]: I0318 10:27:02.847421 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e752ae43-529b-407b-8346-a9eb89990c1f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.391038 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-m6rhx" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.526057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qpfbg" event={"ID":"e752ae43-529b-407b-8346-a9eb89990c1f","Type":"ContainerDied","Data":"a969dc374c564078153562a05b2d6e6f30e62624c3adff3d6bf3cbebc8ffe206"} Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.526817 4733 scope.go:117] "RemoveContainer" containerID="8a6db4a96078cce01cf2e13028e9a6e04a621dfaaea4705c4e07797befc3b96f" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.526144 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qpfbg" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.560517 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.563928 4733 scope.go:117] "RemoveContainer" containerID="d84796ff2be656e89a749a5e50910da1185cb7b35004070858a0b58ba51bc5ea" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.564330 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qpfbg"] Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.590041 4733 scope.go:117] "RemoveContainer" containerID="c39a8c0e7870cb3625474393180c8aa4297d3c1427acc66e1c577dd7840a186c" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.777787 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.778249 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:03 crc kubenswrapper[4733]: I0318 10:27:03.846374 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:04 crc kubenswrapper[4733]: I0318 10:27:04.616323 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:05 crc kubenswrapper[4733]: I0318 10:27:05.191085 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" path="/var/lib/kubelet/pods/e752ae43-529b-407b-8346-a9eb89990c1f/volumes" Mar 18 10:27:07 crc kubenswrapper[4733]: I0318 10:27:07.418748 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:27:07 crc kubenswrapper[4733]: I0318 10:27:07.419141 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5rwz6" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="registry-server" containerID="cri-o://c6d8dc1c0b97ce23301426f81a422e6e5a4b20b07fb86d1870204fc64f69c8a8" gracePeriod=2 Mar 18 10:27:07 crc kubenswrapper[4733]: I0318 10:27:07.568012 4733 generic.go:334] "Generic (PLEG): container finished" podID="c2eed320-8685-4d5c-8793-e506593a33ca" containerID="c6d8dc1c0b97ce23301426f81a422e6e5a4b20b07fb86d1870204fc64f69c8a8" exitCode=0 Mar 18 10:27:07 crc kubenswrapper[4733]: I0318 10:27:07.568085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerDied","Data":"c6d8dc1c0b97ce23301426f81a422e6e5a4b20b07fb86d1870204fc64f69c8a8"} Mar 18 10:27:07 crc kubenswrapper[4733]: I0318 10:27:07.878504 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.029831 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgw6d\" (UniqueName: \"kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d\") pod \"c2eed320-8685-4d5c-8793-e506593a33ca\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.029925 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content\") pod \"c2eed320-8685-4d5c-8793-e506593a33ca\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.029960 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities\") pod \"c2eed320-8685-4d5c-8793-e506593a33ca\" (UID: \"c2eed320-8685-4d5c-8793-e506593a33ca\") " Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.031406 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities" (OuterVolumeSpecName: "utilities") pod "c2eed320-8685-4d5c-8793-e506593a33ca" (UID: "c2eed320-8685-4d5c-8793-e506593a33ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.045458 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d" (OuterVolumeSpecName: "kube-api-access-qgw6d") pod "c2eed320-8685-4d5c-8793-e506593a33ca" (UID: "c2eed320-8685-4d5c-8793-e506593a33ca"). InnerVolumeSpecName "kube-api-access-qgw6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.077177 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2eed320-8685-4d5c-8793-e506593a33ca" (UID: "c2eed320-8685-4d5c-8793-e506593a33ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.132365 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgw6d\" (UniqueName: \"kubernetes.io/projected/c2eed320-8685-4d5c-8793-e506593a33ca-kube-api-access-qgw6d\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.132400 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.132410 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2eed320-8685-4d5c-8793-e506593a33ca-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.579358 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rwz6" event={"ID":"c2eed320-8685-4d5c-8793-e506593a33ca","Type":"ContainerDied","Data":"a61df60801b4e77be52dfd9c59138a93d02124c8a4487a0d508ef12a254c3910"} Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.579517 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rwz6" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.579815 4733 scope.go:117] "RemoveContainer" containerID="c6d8dc1c0b97ce23301426f81a422e6e5a4b20b07fb86d1870204fc64f69c8a8" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.611843 4733 scope.go:117] "RemoveContainer" containerID="c6107d273465cbf0ca33eea0159f3ea649eae43a1b11f06f2154604037aa32c1" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.641497 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.643144 4733 scope.go:117] "RemoveContainer" containerID="e404d64c5fe9561da0e879dd49ffac0a191569c02ab91f0ac47c23b549c8a732" Mar 18 10:27:08 crc kubenswrapper[4733]: I0318 10:27:08.650771 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rwz6"] Mar 18 10:27:09 crc kubenswrapper[4733]: I0318 10:27:09.187947 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" path="/var/lib/kubelet/pods/c2eed320-8685-4d5c-8793-e506593a33ca/volumes" Mar 18 10:27:13 crc kubenswrapper[4733]: I0318 10:27:13.571519 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:27:13 crc kubenswrapper[4733]: I0318 10:27:13.572286 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:27:13 crc kubenswrapper[4733]: I0318 10:27:13.572343 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:27:13 crc kubenswrapper[4733]: I0318 10:27:13.573089 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:27:13 crc kubenswrapper[4733]: I0318 10:27:13.573165 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c" gracePeriod=600 Mar 18 10:27:14 crc kubenswrapper[4733]: I0318 10:27:14.627629 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c" exitCode=0 Mar 18 10:27:14 crc kubenswrapper[4733]: I0318 10:27:14.627722 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c"} Mar 18 10:27:14 crc kubenswrapper[4733]: I0318 10:27:14.628275 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96"} Mar 18 10:27:14 crc kubenswrapper[4733]: I0318 10:27:14.628297 4733 scope.go:117] "RemoveContainer" containerID="bff727181393f1168072f98fbfc5cda5acfb0782a9ae8a688a8335ed7323a527" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.076552 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb"] Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079488 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079508 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079523 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="extract-content" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079531 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="extract-content" Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079543 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="extract-utilities" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079551 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="extract-utilities" Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079564 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="extract-content" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079572 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="extract-content" Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079585 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079592 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: E0318 10:27:18.079600 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="extract-utilities" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079608 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="extract-utilities" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079728 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2eed320-8685-4d5c-8793-e506593a33ca" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.079742 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e752ae43-529b-407b-8346-a9eb89990c1f" containerName="registry-server" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.080802 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.089687 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.099980 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb"] Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.205545 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.205887 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.206133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnrws\" (UniqueName: \"kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.308366 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnrws\" (UniqueName: \"kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.308520 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.308669 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.309645 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.309817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.342935 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnrws\" (UniqueName: \"kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.399522 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:18 crc kubenswrapper[4733]: I0318 10:27:18.758267 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb"] Mar 18 10:27:19 crc kubenswrapper[4733]: I0318 10:27:19.560594 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8v244" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerName="console" containerID="cri-o://bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1" gracePeriod=15 Mar 18 10:27:19 crc kubenswrapper[4733]: I0318 10:27:19.677540 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerID="45f878e96607b8811e5b3070c9def994347e7a1c80f639b1670a63f141ea7cc8" exitCode=0 Mar 18 10:27:19 crc kubenswrapper[4733]: I0318 10:27:19.677589 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" event={"ID":"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6","Type":"ContainerDied","Data":"45f878e96607b8811e5b3070c9def994347e7a1c80f639b1670a63f141ea7cc8"} Mar 18 10:27:19 crc kubenswrapper[4733]: I0318 10:27:19.677618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" event={"ID":"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6","Type":"ContainerStarted","Data":"1be03341496d64c1aaa71d45ded69293eb224d4a64bdfa61810547dec07a2582"} Mar 18 10:27:19 crc kubenswrapper[4733]: I0318 10:27:19.685070 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.066807 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8v244_f27409fc-b6dd-4573-918b-7b30b3635cc7/console/0.log" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.067350 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.249016 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdltm\" (UniqueName: \"kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.249121 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.249165 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.249275 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.250270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config" (OuterVolumeSpecName: "console-config") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.250335 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca" (OuterVolumeSpecName: "service-ca") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.250484 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.251320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.251872 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.251950 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle\") pod \"f27409fc-b6dd-4573-918b-7b30b3635cc7\" (UID: \"f27409fc-b6dd-4573-918b-7b30b3635cc7\") " Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.252510 4733 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.252538 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.252558 4733 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.252538 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.259529 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.260154 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.262022 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm" (OuterVolumeSpecName: "kube-api-access-cdltm") pod "f27409fc-b6dd-4573-918b-7b30b3635cc7" (UID: "f27409fc-b6dd-4573-918b-7b30b3635cc7"). InnerVolumeSpecName "kube-api-access-cdltm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.353968 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdltm\" (UniqueName: \"kubernetes.io/projected/f27409fc-b6dd-4573-918b-7b30b3635cc7-kube-api-access-cdltm\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.354031 4733 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.354051 4733 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f27409fc-b6dd-4573-918b-7b30b3635cc7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.354070 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f27409fc-b6dd-4573-918b-7b30b3635cc7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688441 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8v244_f27409fc-b6dd-4573-918b-7b30b3635cc7/console/0.log" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688519 4733 generic.go:334] "Generic (PLEG): container finished" podID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerID="bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1" exitCode=2 Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8v244" event={"ID":"f27409fc-b6dd-4573-918b-7b30b3635cc7","Type":"ContainerDied","Data":"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1"} Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688606 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8v244" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688634 4733 scope.go:117] "RemoveContainer" containerID="bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.688616 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8v244" event={"ID":"f27409fc-b6dd-4573-918b-7b30b3635cc7","Type":"ContainerDied","Data":"a5e5da5d6249a1112447a42843768f7217f63fd427eb58063240eac26ad5daee"} Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.715359 4733 scope.go:117] "RemoveContainer" containerID="bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1" Mar 18 10:27:20 crc kubenswrapper[4733]: E0318 10:27:20.716162 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1\": container with ID starting with bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1 not found: ID does not exist" containerID="bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.716248 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1"} err="failed to get container status \"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1\": rpc error: code = NotFound desc = could not find container \"bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1\": container with ID starting with bd6ef4d994ae506be5343c7bd62e3c9d5c8d51a521ee2a66c4d08bede745d9e1 not found: ID does not exist" Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.736951 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:27:20 crc kubenswrapper[4733]: I0318 10:27:20.743973 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8v244"] Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.189076 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" path="/var/lib/kubelet/pods/f27409fc-b6dd-4573-918b-7b30b3635cc7/volumes" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.644677 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:21 crc kubenswrapper[4733]: E0318 10:27:21.645611 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerName="console" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.645638 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerName="console" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.646044 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27409fc-b6dd-4573-918b-7b30b3635cc7" containerName="console" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.648249 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.672983 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.701393 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerID="c4d67150099953817c49555f32364536b6d9fd3cb9344b78203a23df29609609" exitCode=0 Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.701455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" event={"ID":"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6","Type":"ContainerDied","Data":"c4d67150099953817c49555f32364536b6d9fd3cb9344b78203a23df29609609"} Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.774927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.775038 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.775255 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv2q6\" (UniqueName: \"kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.876668 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv2q6\" (UniqueName: \"kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.876749 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.876827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.877445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.877520 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.898716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv2q6\" (UniqueName: \"kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6\") pod \"redhat-operators-zcphv\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:21 crc kubenswrapper[4733]: I0318 10:27:21.984784 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.450509 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.714127 4733 generic.go:334] "Generic (PLEG): container finished" podID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerID="f480edb077b6a47bec5706c08fdabad3731b30c4200baa35b38bed5c3ca5beac" exitCode=0 Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.714241 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" event={"ID":"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6","Type":"ContainerDied","Data":"f480edb077b6a47bec5706c08fdabad3731b30c4200baa35b38bed5c3ca5beac"} Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.716992 4733 generic.go:334] "Generic (PLEG): container finished" podID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerID="f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05" exitCode=0 Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.717057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerDied","Data":"f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05"} Mar 18 10:27:22 crc kubenswrapper[4733]: I0318 10:27:22.717094 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerStarted","Data":"1acf1672f2e3aed16d46401b9ad4350dbc89e72639a53dcefec1594e4af7cc4b"} Mar 18 10:27:23 crc kubenswrapper[4733]: I0318 10:27:23.726201 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerStarted","Data":"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214"} Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.046567 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.210970 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle\") pod \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.211734 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util\") pod \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.211783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnrws\" (UniqueName: \"kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws\") pod \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\" (UID: \"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6\") " Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.213156 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle" (OuterVolumeSpecName: "bundle") pod "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" (UID: "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.220282 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws" (OuterVolumeSpecName: "kube-api-access-cnrws") pod "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" (UID: "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6"). InnerVolumeSpecName "kube-api-access-cnrws". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.314162 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.314239 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnrws\" (UniqueName: \"kubernetes.io/projected/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-kube-api-access-cnrws\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.475463 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util" (OuterVolumeSpecName: "util") pod "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" (UID: "2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.517558 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6-util\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.737976 4733 generic.go:334] "Generic (PLEG): container finished" podID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerID="8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214" exitCode=0 Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.738161 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerDied","Data":"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214"} Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.742535 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" event={"ID":"2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6","Type":"ContainerDied","Data":"1be03341496d64c1aaa71d45ded69293eb224d4a64bdfa61810547dec07a2582"} Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.742608 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1be03341496d64c1aaa71d45ded69293eb224d4a64bdfa61810547dec07a2582" Mar 18 10:27:24 crc kubenswrapper[4733]: I0318 10:27:24.742756 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb" Mar 18 10:27:25 crc kubenswrapper[4733]: I0318 10:27:25.755550 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerStarted","Data":"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8"} Mar 18 10:27:25 crc kubenswrapper[4733]: I0318 10:27:25.790982 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zcphv" podStartSLOduration=2.317880858 podStartE2EDuration="4.790955855s" podCreationTimestamp="2026-03-18 10:27:21 +0000 UTC" firstStartedPulling="2026-03-18 10:27:22.71855391 +0000 UTC m=+882.210288235" lastFinishedPulling="2026-03-18 10:27:25.191628867 +0000 UTC m=+884.683363232" observedRunningTime="2026-03-18 10:27:25.782787662 +0000 UTC m=+885.274522047" watchObservedRunningTime="2026-03-18 10:27:25.790955855 +0000 UTC m=+885.282690210" Mar 18 10:27:31 crc kubenswrapper[4733]: I0318 10:27:31.985033 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:31 crc kubenswrapper[4733]: I0318 10:27:31.986822 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:33 crc kubenswrapper[4733]: I0318 10:27:33.040993 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zcphv" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="registry-server" probeResult="failure" output=< Mar 18 10:27:33 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:27:33 crc kubenswrapper[4733]: > Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.633864 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z"] Mar 18 10:27:36 crc kubenswrapper[4733]: E0318 10:27:36.634390 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="pull" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.634404 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="pull" Mar 18 10:27:36 crc kubenswrapper[4733]: E0318 10:27:36.634416 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="util" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.634422 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="util" Mar 18 10:27:36 crc kubenswrapper[4733]: E0318 10:27:36.634436 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="extract" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.634443 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="extract" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.634537 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6" containerName="extract" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.634889 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.637034 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.637456 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.637515 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-ln5pm" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.637565 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.637588 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.668342 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z"] Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.699958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-apiservice-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.700039 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-webhook-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.700143 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frt5k\" (UniqueName: \"kubernetes.io/projected/9731a250-9d78-43e0-bde3-7e769ea43d11-kube-api-access-frt5k\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.800872 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frt5k\" (UniqueName: \"kubernetes.io/projected/9731a250-9d78-43e0-bde3-7e769ea43d11-kube-api-access-frt5k\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.800927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-apiservice-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.800961 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-webhook-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.807951 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-webhook-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.807987 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9731a250-9d78-43e0-bde3-7e769ea43d11-apiservice-cert\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.822916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frt5k\" (UniqueName: \"kubernetes.io/projected/9731a250-9d78-43e0-bde3-7e769ea43d11-kube-api-access-frt5k\") pod \"metallb-operator-controller-manager-5ddc5ff65-jst9z\" (UID: \"9731a250-9d78-43e0-bde3-7e769ea43d11\") " pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.877935 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9"] Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.878629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.880553 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.880761 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xkx8r" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.882454 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.899680 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9"] Mar 18 10:27:36 crc kubenswrapper[4733]: I0318 10:27:36.955823 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.002920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nqfj\" (UniqueName: \"kubernetes.io/projected/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-kube-api-access-7nqfj\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.002984 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-apiservice-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.003046 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-webhook-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.104092 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nqfj\" (UniqueName: \"kubernetes.io/projected/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-kube-api-access-7nqfj\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.104456 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-apiservice-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.104491 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-webhook-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.108092 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-apiservice-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.108287 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-webhook-cert\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.135645 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nqfj\" (UniqueName: \"kubernetes.io/projected/37ecdf54-7bcf-4d33-9cd9-f156974ea7f9-kube-api-access-7nqfj\") pod \"metallb-operator-webhook-server-c99d9f4d6-n5lc9\" (UID: \"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9\") " pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.188830 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z"] Mar 18 10:27:37 crc kubenswrapper[4733]: W0318 10:27:37.192527 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9731a250_9d78_43e0_bde3_7e769ea43d11.slice/crio-93342bba1bffe5603753132a52801bfdc07b439a3f8211da7437bac551bcc01e WatchSource:0}: Error finding container 93342bba1bffe5603753132a52801bfdc07b439a3f8211da7437bac551bcc01e: Status 404 returned error can't find the container with id 93342bba1bffe5603753132a52801bfdc07b439a3f8211da7437bac551bcc01e Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.194624 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.535099 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9"] Mar 18 10:27:37 crc kubenswrapper[4733]: W0318 10:27:37.544852 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37ecdf54_7bcf_4d33_9cd9_f156974ea7f9.slice/crio-7f2887f7ee581e709c23e9fffb49d16eaaf7ff388a25afd577aed07a80b1fe61 WatchSource:0}: Error finding container 7f2887f7ee581e709c23e9fffb49d16eaaf7ff388a25afd577aed07a80b1fe61: Status 404 returned error can't find the container with id 7f2887f7ee581e709c23e9fffb49d16eaaf7ff388a25afd577aed07a80b1fe61 Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.855215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" event={"ID":"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9","Type":"ContainerStarted","Data":"7f2887f7ee581e709c23e9fffb49d16eaaf7ff388a25afd577aed07a80b1fe61"} Mar 18 10:27:37 crc kubenswrapper[4733]: I0318 10:27:37.862823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" event={"ID":"9731a250-9d78-43e0-bde3-7e769ea43d11","Type":"ContainerStarted","Data":"93342bba1bffe5603753132a52801bfdc07b439a3f8211da7437bac551bcc01e"} Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.023932 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.075524 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.262522 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.898348 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" event={"ID":"9731a250-9d78-43e0-bde3-7e769ea43d11","Type":"ContainerStarted","Data":"5ef57c73979a2f29c24bd893efac1018c5c8476eb95c9fd28d2fac812ffd78e5"} Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.898967 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.902553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" event={"ID":"37ecdf54-7bcf-4d33-9cd9-f156974ea7f9","Type":"ContainerStarted","Data":"7dbe890480beca5b6ca7f16045b77235d3e37c02e155e8120b1b7381bfa3e1c0"} Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.902614 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.926636 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" podStartSLOduration=1.8370444620000002 podStartE2EDuration="6.926607843s" podCreationTimestamp="2026-03-18 10:27:36 +0000 UTC" firstStartedPulling="2026-03-18 10:27:37.194743217 +0000 UTC m=+896.686477562" lastFinishedPulling="2026-03-18 10:27:42.284306618 +0000 UTC m=+901.776040943" observedRunningTime="2026-03-18 10:27:42.922888587 +0000 UTC m=+902.414622912" watchObservedRunningTime="2026-03-18 10:27:42.926607843 +0000 UTC m=+902.418342188" Mar 18 10:27:42 crc kubenswrapper[4733]: I0318 10:27:42.961110 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" podStartSLOduration=2.208227574 podStartE2EDuration="6.961091228s" podCreationTimestamp="2026-03-18 10:27:36 +0000 UTC" firstStartedPulling="2026-03-18 10:27:37.547528714 +0000 UTC m=+897.039263029" lastFinishedPulling="2026-03-18 10:27:42.300392358 +0000 UTC m=+901.792126683" observedRunningTime="2026-03-18 10:27:42.957277789 +0000 UTC m=+902.449012124" watchObservedRunningTime="2026-03-18 10:27:42.961091228 +0000 UTC m=+902.452825563" Mar 18 10:27:43 crc kubenswrapper[4733]: I0318 10:27:43.907555 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zcphv" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="registry-server" containerID="cri-o://d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8" gracePeriod=2 Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.305278 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.309504 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities\") pod \"a94fa194-a338-4d4d-9b7a-12440afb4e22\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.309548 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv2q6\" (UniqueName: \"kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6\") pod \"a94fa194-a338-4d4d-9b7a-12440afb4e22\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.309570 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content\") pod \"a94fa194-a338-4d4d-9b7a-12440afb4e22\" (UID: \"a94fa194-a338-4d4d-9b7a-12440afb4e22\") " Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.310302 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities" (OuterVolumeSpecName: "utilities") pod "a94fa194-a338-4d4d-9b7a-12440afb4e22" (UID: "a94fa194-a338-4d4d-9b7a-12440afb4e22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.316044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6" (OuterVolumeSpecName: "kube-api-access-zv2q6") pod "a94fa194-a338-4d4d-9b7a-12440afb4e22" (UID: "a94fa194-a338-4d4d-9b7a-12440afb4e22"). InnerVolumeSpecName "kube-api-access-zv2q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.411333 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv2q6\" (UniqueName: \"kubernetes.io/projected/a94fa194-a338-4d4d-9b7a-12440afb4e22-kube-api-access-zv2q6\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.411367 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.456127 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a94fa194-a338-4d4d-9b7a-12440afb4e22" (UID: "a94fa194-a338-4d4d-9b7a-12440afb4e22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.512285 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a94fa194-a338-4d4d-9b7a-12440afb4e22-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.918475 4733 generic.go:334] "Generic (PLEG): container finished" podID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerID="d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8" exitCode=0 Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.918575 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcphv" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.918578 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerDied","Data":"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8"} Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.919510 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcphv" event={"ID":"a94fa194-a338-4d4d-9b7a-12440afb4e22","Type":"ContainerDied","Data":"1acf1672f2e3aed16d46401b9ad4350dbc89e72639a53dcefec1594e4af7cc4b"} Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.919536 4733 scope.go:117] "RemoveContainer" containerID="d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.946135 4733 scope.go:117] "RemoveContainer" containerID="8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214" Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.960691 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.968832 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zcphv"] Mar 18 10:27:44 crc kubenswrapper[4733]: I0318 10:27:44.997798 4733 scope.go:117] "RemoveContainer" containerID="f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.015961 4733 scope.go:117] "RemoveContainer" containerID="d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8" Mar 18 10:27:45 crc kubenswrapper[4733]: E0318 10:27:45.017626 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8\": container with ID starting with d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8 not found: ID does not exist" containerID="d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.017830 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8"} err="failed to get container status \"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8\": rpc error: code = NotFound desc = could not find container \"d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8\": container with ID starting with d10fef6862efb0197934c1e27e0535679692d0bab9a9c4f0d5d699fb93ec91c8 not found: ID does not exist" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.017972 4733 scope.go:117] "RemoveContainer" containerID="8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214" Mar 18 10:27:45 crc kubenswrapper[4733]: E0318 10:27:45.018645 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214\": container with ID starting with 8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214 not found: ID does not exist" containerID="8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.018718 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214"} err="failed to get container status \"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214\": rpc error: code = NotFound desc = could not find container \"8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214\": container with ID starting with 8fef4e90a34ce3bc76baae5a86a5806e4e83002d16b9733dc19bc2636f351214 not found: ID does not exist" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.018766 4733 scope.go:117] "RemoveContainer" containerID="f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05" Mar 18 10:27:45 crc kubenswrapper[4733]: E0318 10:27:45.019371 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05\": container with ID starting with f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05 not found: ID does not exist" containerID="f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.019454 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05"} err="failed to get container status \"f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05\": rpc error: code = NotFound desc = could not find container \"f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05\": container with ID starting with f9bed651ee1eb90b0878bd6c19ea89ecbdc4aec7e6e5c4b2a087036bf1d47f05 not found: ID does not exist" Mar 18 10:27:45 crc kubenswrapper[4733]: I0318 10:27:45.183261 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" path="/var/lib/kubelet/pods/a94fa194-a338-4d4d-9b7a-12440afb4e22/volumes" Mar 18 10:27:57 crc kubenswrapper[4733]: I0318 10:27:57.205989 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-c99d9f4d6-n5lc9" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.152254 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563828-zczv7"] Mar 18 10:28:00 crc kubenswrapper[4733]: E0318 10:28:00.152775 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="registry-server" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.152788 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="registry-server" Mar 18 10:28:00 crc kubenswrapper[4733]: E0318 10:28:00.152805 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="extract-content" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.152813 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="extract-content" Mar 18 10:28:00 crc kubenswrapper[4733]: E0318 10:28:00.152826 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="extract-utilities" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.152835 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="extract-utilities" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.152973 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a94fa194-a338-4d4d-9b7a-12440afb4e22" containerName="registry-server" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.153469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.156559 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.157109 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.158006 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.172863 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563828-zczv7"] Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.346220 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smzsq\" (UniqueName: \"kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq\") pod \"auto-csr-approver-29563828-zczv7\" (UID: \"68574d72-725d-48c2-b645-bd83dcccbf80\") " pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.447246 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smzsq\" (UniqueName: \"kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq\") pod \"auto-csr-approver-29563828-zczv7\" (UID: \"68574d72-725d-48c2-b645-bd83dcccbf80\") " pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.484161 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smzsq\" (UniqueName: \"kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq\") pod \"auto-csr-approver-29563828-zczv7\" (UID: \"68574d72-725d-48c2-b645-bd83dcccbf80\") " pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.484684 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:00 crc kubenswrapper[4733]: I0318 10:28:00.779876 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563828-zczv7"] Mar 18 10:28:01 crc kubenswrapper[4733]: I0318 10:28:01.034295 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563828-zczv7" event={"ID":"68574d72-725d-48c2-b645-bd83dcccbf80","Type":"ContainerStarted","Data":"5b7e6b80bc4fb02a1f7afcf04f2e7e7266bf3a39aab1842bdbd19ce39cd1c153"} Mar 18 10:28:03 crc kubenswrapper[4733]: I0318 10:28:03.054551 4733 generic.go:334] "Generic (PLEG): container finished" podID="68574d72-725d-48c2-b645-bd83dcccbf80" containerID="f4a3549ea82cce03bd994263d641938a407bdfdc2f86792bccee0b653493614d" exitCode=0 Mar 18 10:28:03 crc kubenswrapper[4733]: I0318 10:28:03.054645 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563828-zczv7" event={"ID":"68574d72-725d-48c2-b645-bd83dcccbf80","Type":"ContainerDied","Data":"f4a3549ea82cce03bd994263d641938a407bdfdc2f86792bccee0b653493614d"} Mar 18 10:28:04 crc kubenswrapper[4733]: I0318 10:28:04.306954 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:04 crc kubenswrapper[4733]: I0318 10:28:04.507787 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smzsq\" (UniqueName: \"kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq\") pod \"68574d72-725d-48c2-b645-bd83dcccbf80\" (UID: \"68574d72-725d-48c2-b645-bd83dcccbf80\") " Mar 18 10:28:04 crc kubenswrapper[4733]: I0318 10:28:04.516365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq" (OuterVolumeSpecName: "kube-api-access-smzsq") pod "68574d72-725d-48c2-b645-bd83dcccbf80" (UID: "68574d72-725d-48c2-b645-bd83dcccbf80"). InnerVolumeSpecName "kube-api-access-smzsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:28:04 crc kubenswrapper[4733]: I0318 10:28:04.609625 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smzsq\" (UniqueName: \"kubernetes.io/projected/68574d72-725d-48c2-b645-bd83dcccbf80-kube-api-access-smzsq\") on node \"crc\" DevicePath \"\"" Mar 18 10:28:05 crc kubenswrapper[4733]: I0318 10:28:05.071737 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563828-zczv7" event={"ID":"68574d72-725d-48c2-b645-bd83dcccbf80","Type":"ContainerDied","Data":"5b7e6b80bc4fb02a1f7afcf04f2e7e7266bf3a39aab1842bdbd19ce39cd1c153"} Mar 18 10:28:05 crc kubenswrapper[4733]: I0318 10:28:05.071816 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b7e6b80bc4fb02a1f7afcf04f2e7e7266bf3a39aab1842bdbd19ce39cd1c153" Mar 18 10:28:05 crc kubenswrapper[4733]: I0318 10:28:05.071828 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563828-zczv7" Mar 18 10:28:05 crc kubenswrapper[4733]: I0318 10:28:05.375500 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563822-4fvb8"] Mar 18 10:28:05 crc kubenswrapper[4733]: I0318 10:28:05.382542 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563822-4fvb8"] Mar 18 10:28:07 crc kubenswrapper[4733]: I0318 10:28:07.189486 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f93d05a-41f2-4422-88aa-9dfddb13191f" path="/var/lib/kubelet/pods/3f93d05a-41f2-4422-88aa-9dfddb13191f/volumes" Mar 18 10:28:16 crc kubenswrapper[4733]: I0318 10:28:16.961275 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5ddc5ff65-jst9z" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.744795 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pc5zz"] Mar 18 10:28:17 crc kubenswrapper[4733]: E0318 10:28:17.745069 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68574d72-725d-48c2-b645-bd83dcccbf80" containerName="oc" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.745092 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="68574d72-725d-48c2-b645-bd83dcccbf80" containerName="oc" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.745253 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="68574d72-725d-48c2-b645-bd83dcccbf80" containerName="oc" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.747537 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.748990 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.751505 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.751816 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5nckh" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.751883 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg"] Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.752935 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.755447 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.761833 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg"] Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.829905 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zg5cv"] Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.830965 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zg5cv" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.833736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.833768 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.833788 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.834460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-cbdb8" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.846388 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-zsljc"] Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.847541 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.849492 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.853988 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-zsljc"] Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp57k\" (UniqueName: \"kubernetes.io/projected/03476444-8ff8-4b1e-bcbc-ee654241370b-kube-api-access-lp57k\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03476444-8ff8-4b1e-bcbc-ee654241370b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900608 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-startup\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900627 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-conf\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900662 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-sockets\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900704 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics-certs\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900726 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvj9w\" (UniqueName: \"kubernetes.io/projected/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-kube-api-access-pvj9w\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:17 crc kubenswrapper[4733]: I0318 10:28:17.900812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-reloader\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002643 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-startup\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-conf\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002688 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zkdn\" (UniqueName: \"kubernetes.io/projected/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-kube-api-access-9zkdn\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/101c5687-bebd-449f-94c8-03077bf596d0-metallb-excludel2\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002792 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84g5w\" (UniqueName: \"kubernetes.io/projected/101c5687-bebd-449f-94c8-03077bf596d0-kube-api-access-84g5w\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-sockets\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.002866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics-certs\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003148 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-conf\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003247 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-sockets\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003299 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-metrics-certs\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003343 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-metrics-certs\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvj9w\" (UniqueName: \"kubernetes.io/projected/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-kube-api-access-pvj9w\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003411 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-reloader\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003616 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-frr-startup\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-reloader\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003430 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp57k\" (UniqueName: \"kubernetes.io/projected/03476444-8ff8-4b1e-bcbc-ee654241370b-kube-api-access-lp57k\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003695 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-cert\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.003950 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.004010 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03476444-8ff8-4b1e-bcbc-ee654241370b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.008543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-metrics-certs\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.016358 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03476444-8ff8-4b1e-bcbc-ee654241370b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.020545 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp57k\" (UniqueName: \"kubernetes.io/projected/03476444-8ff8-4b1e-bcbc-ee654241370b-kube-api-access-lp57k\") pod \"frr-k8s-webhook-server-bcc4b6f68-dr9dg\" (UID: \"03476444-8ff8-4b1e-bcbc-ee654241370b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.023642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvj9w\" (UniqueName: \"kubernetes.io/projected/4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e-kube-api-access-pvj9w\") pod \"frr-k8s-pc5zz\" (UID: \"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e\") " pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.066618 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.074491 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105376 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zkdn\" (UniqueName: \"kubernetes.io/projected/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-kube-api-access-9zkdn\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105488 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/101c5687-bebd-449f-94c8-03077bf596d0-metallb-excludel2\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105511 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84g5w\" (UniqueName: \"kubernetes.io/projected/101c5687-bebd-449f-94c8-03077bf596d0-kube-api-access-84g5w\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: E0318 10:28:18.105516 4733 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105553 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-metrics-certs\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: E0318 10:28:18.105589 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist podName:101c5687-bebd-449f-94c8-03077bf596d0 nodeName:}" failed. No retries permitted until 2026-03-18 10:28:18.605567199 +0000 UTC m=+938.097301524 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist") pod "speaker-zg5cv" (UID: "101c5687-bebd-449f-94c8-03077bf596d0") : secret "metallb-memberlist" not found Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105613 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-metrics-certs\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.105716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-cert\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.106754 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/101c5687-bebd-449f-94c8-03077bf596d0-metallb-excludel2\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.107984 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.109479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-metrics-certs\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.112519 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-metrics-certs\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.120642 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-cert\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.133346 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84g5w\" (UniqueName: \"kubernetes.io/projected/101c5687-bebd-449f-94c8-03077bf596d0-kube-api-access-84g5w\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.140999 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zkdn\" (UniqueName: \"kubernetes.io/projected/7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0-kube-api-access-9zkdn\") pod \"controller-7bb4cc7c98-zsljc\" (UID: \"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0\") " pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.159951 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.390809 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-zsljc"] Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.543499 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg"] Mar 18 10:28:18 crc kubenswrapper[4733]: W0318 10:28:18.553517 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03476444_8ff8_4b1e_bcbc_ee654241370b.slice/crio-2e377ff8591027d484d966029946b79e8ff3f7ceac70bb72e86775a2f0b2d378 WatchSource:0}: Error finding container 2e377ff8591027d484d966029946b79e8ff3f7ceac70bb72e86775a2f0b2d378: Status 404 returned error can't find the container with id 2e377ff8591027d484d966029946b79e8ff3f7ceac70bb72e86775a2f0b2d378 Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.610716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.616386 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/101c5687-bebd-449f-94c8-03077bf596d0-memberlist\") pod \"speaker-zg5cv\" (UID: \"101c5687-bebd-449f-94c8-03077bf596d0\") " pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: I0318 10:28:18.743283 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zg5cv" Mar 18 10:28:18 crc kubenswrapper[4733]: W0318 10:28:18.767720 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod101c5687_bebd_449f_94c8_03077bf596d0.slice/crio-361d42d5a7326ad4e5eae10ddda3f7b97c0199c2d631383b5fa2146dfd825fdd WatchSource:0}: Error finding container 361d42d5a7326ad4e5eae10ddda3f7b97c0199c2d631383b5fa2146dfd825fdd: Status 404 returned error can't find the container with id 361d42d5a7326ad4e5eae10ddda3f7b97c0199c2d631383b5fa2146dfd825fdd Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.212845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-zsljc" event={"ID":"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0","Type":"ContainerStarted","Data":"ba7b307fc134f9369d0894115c979129b8eb46fcc62b5c869e057899a1e2b1cf"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.213288 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.213309 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-zsljc" event={"ID":"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0","Type":"ContainerStarted","Data":"73ba0541c10f5b0a50e8ce47bb8a17737bc73f260ca54632bdd84ee798bebe1f"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.213322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-zsljc" event={"ID":"7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0","Type":"ContainerStarted","Data":"115c888652219a3db8c1591a5cf20ced90238786397975c66c708d7deb9caf63"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.214857 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zg5cv" event={"ID":"101c5687-bebd-449f-94c8-03077bf596d0","Type":"ContainerStarted","Data":"492daea53269b124abeba4fabdb1687ec7eb9f9c7b9c930a5b934ddd7bd1690a"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.214891 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zg5cv" event={"ID":"101c5687-bebd-449f-94c8-03077bf596d0","Type":"ContainerStarted","Data":"361d42d5a7326ad4e5eae10ddda3f7b97c0199c2d631383b5fa2146dfd825fdd"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.215939 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"d3ce1c0fa6046f7621200f51af9eca75a94739beedb1c458b841612084cce14d"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.216970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" event={"ID":"03476444-8ff8-4b1e-bcbc-ee654241370b","Type":"ContainerStarted","Data":"2e377ff8591027d484d966029946b79e8ff3f7ceac70bb72e86775a2f0b2d378"} Mar 18 10:28:19 crc kubenswrapper[4733]: I0318 10:28:19.235009 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-zsljc" podStartSLOduration=2.234987441 podStartE2EDuration="2.234987441s" podCreationTimestamp="2026-03-18 10:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:28:19.230663368 +0000 UTC m=+938.722397713" watchObservedRunningTime="2026-03-18 10:28:19.234987441 +0000 UTC m=+938.726721766" Mar 18 10:28:20 crc kubenswrapper[4733]: I0318 10:28:20.227833 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zg5cv" event={"ID":"101c5687-bebd-449f-94c8-03077bf596d0","Type":"ContainerStarted","Data":"8b1c9943dcbc575fc8ba1fcceaa064a66e93c0d1b38dd42aa5436b6fff4c0b33"} Mar 18 10:28:20 crc kubenswrapper[4733]: I0318 10:28:20.246838 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zg5cv" podStartSLOduration=3.246824036 podStartE2EDuration="3.246824036s" podCreationTimestamp="2026-03-18 10:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:28:20.244866141 +0000 UTC m=+939.736600466" watchObservedRunningTime="2026-03-18 10:28:20.246824036 +0000 UTC m=+939.738558361" Mar 18 10:28:21 crc kubenswrapper[4733]: I0318 10:28:21.232276 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zg5cv" Mar 18 10:28:26 crc kubenswrapper[4733]: I0318 10:28:26.270113 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e" containerID="77592e00ae10c5cb3abbce4b303f094fcb801175e4f65997fe669e8b2217f4a0" exitCode=0 Mar 18 10:28:26 crc kubenswrapper[4733]: I0318 10:28:26.270251 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerDied","Data":"77592e00ae10c5cb3abbce4b303f094fcb801175e4f65997fe669e8b2217f4a0"} Mar 18 10:28:26 crc kubenswrapper[4733]: I0318 10:28:26.274977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" event={"ID":"03476444-8ff8-4b1e-bcbc-ee654241370b","Type":"ContainerStarted","Data":"15e8ccb7c43f4feaba91b540617a0850e2756a1b17a1f22655bbc0aab2ef0119"} Mar 18 10:28:26 crc kubenswrapper[4733]: I0318 10:28:26.275215 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:26 crc kubenswrapper[4733]: I0318 10:28:26.335621 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" podStartSLOduration=1.930053147 podStartE2EDuration="9.335594652s" podCreationTimestamp="2026-03-18 10:28:17 +0000 UTC" firstStartedPulling="2026-03-18 10:28:18.556155486 +0000 UTC m=+938.047889831" lastFinishedPulling="2026-03-18 10:28:25.961697011 +0000 UTC m=+945.453431336" observedRunningTime="2026-03-18 10:28:26.326412982 +0000 UTC m=+945.818147347" watchObservedRunningTime="2026-03-18 10:28:26.335594652 +0000 UTC m=+945.827329007" Mar 18 10:28:27 crc kubenswrapper[4733]: I0318 10:28:27.285684 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e" containerID="19ffeb7c6c69ef8972961611d99d630a3ffff4ad8ae9c2bdcefdae455d13de4e" exitCode=0 Mar 18 10:28:27 crc kubenswrapper[4733]: I0318 10:28:27.285763 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerDied","Data":"19ffeb7c6c69ef8972961611d99d630a3ffff4ad8ae9c2bdcefdae455d13de4e"} Mar 18 10:28:28 crc kubenswrapper[4733]: I0318 10:28:28.165488 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-zsljc" Mar 18 10:28:28 crc kubenswrapper[4733]: I0318 10:28:28.308895 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e" containerID="8f29496ea299a3c1c374b7f0ca06a25badcf18e2cc3b29bdd493053b8730b4c7" exitCode=0 Mar 18 10:28:28 crc kubenswrapper[4733]: I0318 10:28:28.308939 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerDied","Data":"8f29496ea299a3c1c374b7f0ca06a25badcf18e2cc3b29bdd493053b8730b4c7"} Mar 18 10:28:29 crc kubenswrapper[4733]: I0318 10:28:29.324552 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"b1c58c746b575489046dabc4b9328bd430507d943f24275c2f719c3f76f7aae8"} Mar 18 10:28:29 crc kubenswrapper[4733]: I0318 10:28:29.324895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"e474f08af12ff42635fc4df678c01fa724e980a724fc8145d689081cc137a0f0"} Mar 18 10:28:29 crc kubenswrapper[4733]: I0318 10:28:29.324906 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"89057b4cfb461860fbc73703161e75c7cc575a895caae65b565138321b2f5dbd"} Mar 18 10:28:29 crc kubenswrapper[4733]: I0318 10:28:29.324915 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"417383e21e4a51247d15e7ba43740fe7b14bcc54788cb95fc208160cf62e42b7"} Mar 18 10:28:29 crc kubenswrapper[4733]: I0318 10:28:29.324923 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"9a6288e6c81f2c40bc1cb5404e34f2002e68018920483c192b0523532e3f0a05"} Mar 18 10:28:30 crc kubenswrapper[4733]: I0318 10:28:30.338107 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pc5zz" event={"ID":"4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e","Type":"ContainerStarted","Data":"ecdc9b2404abc2ca69731d8334d73462bea7eaeba4cb929a36168096f466678a"} Mar 18 10:28:30 crc kubenswrapper[4733]: I0318 10:28:30.339062 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:30 crc kubenswrapper[4733]: I0318 10:28:30.373625 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pc5zz" podStartSLOduration=5.651094269 podStartE2EDuration="13.373600419s" podCreationTimestamp="2026-03-18 10:28:17 +0000 UTC" firstStartedPulling="2026-03-18 10:28:18.250307247 +0000 UTC m=+937.742041572" lastFinishedPulling="2026-03-18 10:28:25.972813397 +0000 UTC m=+945.464547722" observedRunningTime="2026-03-18 10:28:30.371224052 +0000 UTC m=+949.862958397" watchObservedRunningTime="2026-03-18 10:28:30.373600419 +0000 UTC m=+949.865334754" Mar 18 10:28:33 crc kubenswrapper[4733]: I0318 10:28:33.067120 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:33 crc kubenswrapper[4733]: I0318 10:28:33.123814 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:38 crc kubenswrapper[4733]: I0318 10:28:38.073137 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pc5zz" Mar 18 10:28:38 crc kubenswrapper[4733]: I0318 10:28:38.079079 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" Mar 18 10:28:38 crc kubenswrapper[4733]: I0318 10:28:38.750605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zg5cv" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.091801 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-g2m9r"] Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.094418 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.099917 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8s8qq" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.100866 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.100872 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.118898 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g2m9r"] Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.199519 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqf9r\" (UniqueName: \"kubernetes.io/projected/45605961-e7c2-4bd3-a670-d8541124408a-kube-api-access-hqf9r\") pod \"openstack-operator-index-g2m9r\" (UID: \"45605961-e7c2-4bd3-a670-d8541124408a\") " pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.300647 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqf9r\" (UniqueName: \"kubernetes.io/projected/45605961-e7c2-4bd3-a670-d8541124408a-kube-api-access-hqf9r\") pod \"openstack-operator-index-g2m9r\" (UID: \"45605961-e7c2-4bd3-a670-d8541124408a\") " pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.318810 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqf9r\" (UniqueName: \"kubernetes.io/projected/45605961-e7c2-4bd3-a670-d8541124408a-kube-api-access-hqf9r\") pod \"openstack-operator-index-g2m9r\" (UID: \"45605961-e7c2-4bd3-a670-d8541124408a\") " pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.419049 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:45 crc kubenswrapper[4733]: I0318 10:28:45.672802 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g2m9r"] Mar 18 10:28:46 crc kubenswrapper[4733]: I0318 10:28:46.492056 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g2m9r" event={"ID":"45605961-e7c2-4bd3-a670-d8541124408a","Type":"ContainerStarted","Data":"bf248de56e76050e45ec04123a6cc0d376a854e1d370cac6f0db486c1ee041fb"} Mar 18 10:28:49 crc kubenswrapper[4733]: I0318 10:28:49.524363 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g2m9r" event={"ID":"45605961-e7c2-4bd3-a670-d8541124408a","Type":"ContainerStarted","Data":"df46845da741ad71124c2ae43e10c583672f7a52bd0107a068c1058d79eaf580"} Mar 18 10:28:49 crc kubenswrapper[4733]: I0318 10:28:49.548175 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-g2m9r" podStartSLOduration=1.752075413 podStartE2EDuration="4.548149963s" podCreationTimestamp="2026-03-18 10:28:45 +0000 UTC" firstStartedPulling="2026-03-18 10:28:45.681947573 +0000 UTC m=+965.173681908" lastFinishedPulling="2026-03-18 10:28:48.478022093 +0000 UTC m=+967.969756458" observedRunningTime="2026-03-18 10:28:49.544441488 +0000 UTC m=+969.036175843" watchObservedRunningTime="2026-03-18 10:28:49.548149963 +0000 UTC m=+969.039884328" Mar 18 10:28:54 crc kubenswrapper[4733]: I0318 10:28:54.837741 4733 scope.go:117] "RemoveContainer" containerID="36c7a80bc1a34092c9183dbd958b5c05ea904377be8cffacb7112a1b4663e6a6" Mar 18 10:28:55 crc kubenswrapper[4733]: I0318 10:28:55.420491 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:55 crc kubenswrapper[4733]: I0318 10:28:55.422165 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:55 crc kubenswrapper[4733]: I0318 10:28:55.466512 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:55 crc kubenswrapper[4733]: I0318 10:28:55.610611 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-g2m9r" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.139691 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67"] Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.142231 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.145349 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wwmnn" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.158658 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67"] Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.170941 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j47dh\" (UniqueName: \"kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.171737 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.171975 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.273640 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.273814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j47dh\" (UniqueName: \"kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.273872 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.274799 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.274821 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.297384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j47dh\" (UniqueName: \"kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh\") pod \"77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:57 crc kubenswrapper[4733]: I0318 10:28:57.478684 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:28:58 crc kubenswrapper[4733]: I0318 10:28:58.006308 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67"] Mar 18 10:28:58 crc kubenswrapper[4733]: I0318 10:28:58.596002 4733 generic.go:334] "Generic (PLEG): container finished" podID="53c111d7-ea42-4913-b378-ec44062b0691" containerID="0db45fe63d0f98cced7909596d8d0f8df1bf9f9dfa3875dad551f78b45821259" exitCode=0 Mar 18 10:28:58 crc kubenswrapper[4733]: I0318 10:28:58.596378 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" event={"ID":"53c111d7-ea42-4913-b378-ec44062b0691","Type":"ContainerDied","Data":"0db45fe63d0f98cced7909596d8d0f8df1bf9f9dfa3875dad551f78b45821259"} Mar 18 10:28:58 crc kubenswrapper[4733]: I0318 10:28:58.596417 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" event={"ID":"53c111d7-ea42-4913-b378-ec44062b0691","Type":"ContainerStarted","Data":"ff66937e27c0a955eaec5cd058090359982cd344502efa748ae95de95e350359"} Mar 18 10:28:59 crc kubenswrapper[4733]: I0318 10:28:59.609165 4733 generic.go:334] "Generic (PLEG): container finished" podID="53c111d7-ea42-4913-b378-ec44062b0691" containerID="49ad44c2d5b98fc6566f4f4bfdc1f9289d943440590ef96e4199d07928e7605a" exitCode=0 Mar 18 10:28:59 crc kubenswrapper[4733]: I0318 10:28:59.609250 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" event={"ID":"53c111d7-ea42-4913-b378-ec44062b0691","Type":"ContainerDied","Data":"49ad44c2d5b98fc6566f4f4bfdc1f9289d943440590ef96e4199d07928e7605a"} Mar 18 10:29:00 crc kubenswrapper[4733]: I0318 10:29:00.621300 4733 generic.go:334] "Generic (PLEG): container finished" podID="53c111d7-ea42-4913-b378-ec44062b0691" containerID="49b5d86705056140c0d3154ecb828d2e235cb59eee079edb114bc8df64c3ff85" exitCode=0 Mar 18 10:29:00 crc kubenswrapper[4733]: I0318 10:29:00.621423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" event={"ID":"53c111d7-ea42-4913-b378-ec44062b0691","Type":"ContainerDied","Data":"49b5d86705056140c0d3154ecb828d2e235cb59eee079edb114bc8df64c3ff85"} Mar 18 10:29:01 crc kubenswrapper[4733]: I0318 10:29:01.953437 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.055934 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j47dh\" (UniqueName: \"kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh\") pod \"53c111d7-ea42-4913-b378-ec44062b0691\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.056026 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle\") pod \"53c111d7-ea42-4913-b378-ec44062b0691\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.056102 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util\") pod \"53c111d7-ea42-4913-b378-ec44062b0691\" (UID: \"53c111d7-ea42-4913-b378-ec44062b0691\") " Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.057136 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle" (OuterVolumeSpecName: "bundle") pod "53c111d7-ea42-4913-b378-ec44062b0691" (UID: "53c111d7-ea42-4913-b378-ec44062b0691"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.065878 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh" (OuterVolumeSpecName: "kube-api-access-j47dh") pod "53c111d7-ea42-4913-b378-ec44062b0691" (UID: "53c111d7-ea42-4913-b378-ec44062b0691"). InnerVolumeSpecName "kube-api-access-j47dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.085826 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util" (OuterVolumeSpecName: "util") pod "53c111d7-ea42-4913-b378-ec44062b0691" (UID: "53c111d7-ea42-4913-b378-ec44062b0691"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.158052 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j47dh\" (UniqueName: \"kubernetes.io/projected/53c111d7-ea42-4913-b378-ec44062b0691-kube-api-access-j47dh\") on node \"crc\" DevicePath \"\"" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.158106 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.158124 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53c111d7-ea42-4913-b378-ec44062b0691-util\") on node \"crc\" DevicePath \"\"" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.644326 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" event={"ID":"53c111d7-ea42-4913-b378-ec44062b0691","Type":"ContainerDied","Data":"ff66937e27c0a955eaec5cd058090359982cd344502efa748ae95de95e350359"} Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.644361 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff66937e27c0a955eaec5cd058090359982cd344502efa748ae95de95e350359" Mar 18 10:29:02 crc kubenswrapper[4733]: I0318 10:29:02.644510 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.448668 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4"] Mar 18 10:29:06 crc kubenswrapper[4733]: E0318 10:29:06.449511 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="extract" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.449525 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="extract" Mar 18 10:29:06 crc kubenswrapper[4733]: E0318 10:29:06.449549 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="util" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.449557 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="util" Mar 18 10:29:06 crc kubenswrapper[4733]: E0318 10:29:06.449568 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="pull" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.449576 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="pull" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.449692 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c111d7-ea42-4913-b378-ec44062b0691" containerName="extract" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.450127 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.452448 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-7kkxj" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.527686 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kh4c\" (UniqueName: \"kubernetes.io/projected/d1b10458-2335-4b46-9f63-c8a005096ff7-kube-api-access-2kh4c\") pod \"openstack-operator-controller-init-579f7bfb88-sfsb4\" (UID: \"d1b10458-2335-4b46-9f63-c8a005096ff7\") " pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.544194 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4"] Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.629109 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kh4c\" (UniqueName: \"kubernetes.io/projected/d1b10458-2335-4b46-9f63-c8a005096ff7-kube-api-access-2kh4c\") pod \"openstack-operator-controller-init-579f7bfb88-sfsb4\" (UID: \"d1b10458-2335-4b46-9f63-c8a005096ff7\") " pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.650324 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kh4c\" (UniqueName: \"kubernetes.io/projected/d1b10458-2335-4b46-9f63-c8a005096ff7-kube-api-access-2kh4c\") pod \"openstack-operator-controller-init-579f7bfb88-sfsb4\" (UID: \"d1b10458-2335-4b46-9f63-c8a005096ff7\") " pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:06 crc kubenswrapper[4733]: I0318 10:29:06.830616 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:07 crc kubenswrapper[4733]: W0318 10:29:07.186079 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b10458_2335_4b46_9f63_c8a005096ff7.slice/crio-457e41d42cd11946e9089bf9f569c743a3e1481e1a8619f952166b14fdb3f696 WatchSource:0}: Error finding container 457e41d42cd11946e9089bf9f569c743a3e1481e1a8619f952166b14fdb3f696: Status 404 returned error can't find the container with id 457e41d42cd11946e9089bf9f569c743a3e1481e1a8619f952166b14fdb3f696 Mar 18 10:29:07 crc kubenswrapper[4733]: I0318 10:29:07.187602 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4"] Mar 18 10:29:07 crc kubenswrapper[4733]: I0318 10:29:07.678830 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" event={"ID":"d1b10458-2335-4b46-9f63-c8a005096ff7","Type":"ContainerStarted","Data":"457e41d42cd11946e9089bf9f569c743a3e1481e1a8619f952166b14fdb3f696"} Mar 18 10:29:11 crc kubenswrapper[4733]: I0318 10:29:11.707646 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" event={"ID":"d1b10458-2335-4b46-9f63-c8a005096ff7","Type":"ContainerStarted","Data":"08e673cd0aec30d6831dfb6c5a4e7caae3e8dde0cd321bb81c38a42ca8bdceb2"} Mar 18 10:29:11 crc kubenswrapper[4733]: I0318 10:29:11.708338 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:11 crc kubenswrapper[4733]: I0318 10:29:11.730818 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" podStartSLOduration=2.188099076 podStartE2EDuration="5.730801277s" podCreationTimestamp="2026-03-18 10:29:06 +0000 UTC" firstStartedPulling="2026-03-18 10:29:07.18751695 +0000 UTC m=+986.679251275" lastFinishedPulling="2026-03-18 10:29:10.730219151 +0000 UTC m=+990.221953476" observedRunningTime="2026-03-18 10:29:11.729735887 +0000 UTC m=+991.221470202" watchObservedRunningTime="2026-03-18 10:29:11.730801277 +0000 UTC m=+991.222535602" Mar 18 10:29:13 crc kubenswrapper[4733]: I0318 10:29:13.571085 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:29:13 crc kubenswrapper[4733]: I0318 10:29:13.571614 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:29:16 crc kubenswrapper[4733]: I0318 10:29:16.833877 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-579f7bfb88-sfsb4" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.172333 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.174301 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.176775 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vnvsj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.180146 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.181100 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.188279 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mh8jr" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.198493 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.201625 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrfg8\" (UniqueName: \"kubernetes.io/projected/8fe910c4-798b-4381-a71d-697459f7f79a-kube-api-access-lrfg8\") pod \"cinder-operator-controller-manager-8d58dc466-v6zxn\" (UID: \"8fe910c4-798b-4381-a71d-697459f7f79a\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.201722 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsmw6\" (UniqueName: \"kubernetes.io/projected/0fb2ba68-fa0f-4483-afdf-2eb381c54320-kube-api-access-fsmw6\") pod \"barbican-operator-controller-manager-59bc569d95-sfv8v\" (UID: \"0fb2ba68-fa0f-4483-afdf-2eb381c54320\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.205282 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-t8796"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.206412 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.210876 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-ks7mv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.216282 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.221307 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.222087 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.224850 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p5z7d" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.250156 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-t8796"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.304665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsmw6\" (UniqueName: \"kubernetes.io/projected/0fb2ba68-fa0f-4483-afdf-2eb381c54320-kube-api-access-fsmw6\") pod \"barbican-operator-controller-manager-59bc569d95-sfv8v\" (UID: \"0fb2ba68-fa0f-4483-afdf-2eb381c54320\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.304764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhzwx\" (UniqueName: \"kubernetes.io/projected/bc0e28fc-cff0-4c39-8073-61d5d6481866-kube-api-access-nhzwx\") pod \"glance-operator-controller-manager-79df6bcc97-ljvrt\" (UID: \"bc0e28fc-cff0-4c39-8073-61d5d6481866\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.304824 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv8xv\" (UniqueName: \"kubernetes.io/projected/748f4855-3978-4ecd-805e-0fee34ce0094-kube-api-access-pv8xv\") pod \"designate-operator-controller-manager-588d4d986b-t8796\" (UID: \"748f4855-3978-4ecd-805e-0fee34ce0094\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.305098 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrfg8\" (UniqueName: \"kubernetes.io/projected/8fe910c4-798b-4381-a71d-697459f7f79a-kube-api-access-lrfg8\") pod \"cinder-operator-controller-manager-8d58dc466-v6zxn\" (UID: \"8fe910c4-798b-4381-a71d-697459f7f79a\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.308295 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.318382 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.319463 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.326295 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-h62lg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.339730 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.340704 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.343869 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsmw6\" (UniqueName: \"kubernetes.io/projected/0fb2ba68-fa0f-4483-afdf-2eb381c54320-kube-api-access-fsmw6\") pod \"barbican-operator-controller-manager-59bc569d95-sfv8v\" (UID: \"0fb2ba68-fa0f-4483-afdf-2eb381c54320\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.346505 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bjq7m" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.346836 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrfg8\" (UniqueName: \"kubernetes.io/projected/8fe910c4-798b-4381-a71d-697459f7f79a-kube-api-access-lrfg8\") pod \"cinder-operator-controller-manager-8d58dc466-v6zxn\" (UID: \"8fe910c4-798b-4381-a71d-697459f7f79a\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.362882 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.364091 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.365684 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.366058 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vp5bk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.370965 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.382372 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.383262 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.387941 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-fd55l" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.407105 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.407461 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhzwx\" (UniqueName: \"kubernetes.io/projected/bc0e28fc-cff0-4c39-8073-61d5d6481866-kube-api-access-nhzwx\") pod \"glance-operator-controller-manager-79df6bcc97-ljvrt\" (UID: \"bc0e28fc-cff0-4c39-8073-61d5d6481866\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.407844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv8xv\" (UniqueName: \"kubernetes.io/projected/748f4855-3978-4ecd-805e-0fee34ce0094-kube-api-access-pv8xv\") pod \"designate-operator-controller-manager-588d4d986b-t8796\" (UID: \"748f4855-3978-4ecd-805e-0fee34ce0094\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.425065 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.451244 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.459639 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-chmbd"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.460452 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.460796 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv8xv\" (UniqueName: \"kubernetes.io/projected/748f4855-3978-4ecd-805e-0fee34ce0094-kube-api-access-pv8xv\") pod \"designate-operator-controller-manager-588d4d986b-t8796\" (UID: \"748f4855-3978-4ecd-805e-0fee34ce0094\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.463395 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-28sd9" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.464088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhzwx\" (UniqueName: \"kubernetes.io/projected/bc0e28fc-cff0-4c39-8073-61d5d6481866-kube-api-access-nhzwx\") pod \"glance-operator-controller-manager-79df6bcc97-ljvrt\" (UID: \"bc0e28fc-cff0-4c39-8073-61d5d6481866\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.470102 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.471054 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.472039 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.473493 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.475069 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-7h96r" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.478597 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4n8bg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.494470 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.502177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.503407 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.508508 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-chmbd"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509508 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktf2p\" (UniqueName: \"kubernetes.io/projected/79dfdcde-0538-4777-959e-1daf2b6263de-kube-api-access-ktf2p\") pod \"mariadb-operator-controller-manager-67ccfc9778-4xzlc\" (UID: \"79dfdcde-0538-4777-959e-1daf2b6263de\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnppd\" (UniqueName: \"kubernetes.io/projected/651c7dd5-3adc-48b4-b579-309258aa3735-kube-api-access-pnppd\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509550 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2plph\" (UniqueName: \"kubernetes.io/projected/bd5ae902-d036-4e52-983d-aa3e1a86dca8-kube-api-access-2plph\") pod \"heat-operator-controller-manager-67dd5f86f5-cxlns\" (UID: \"bd5ae902-d036-4e52-983d-aa3e1a86dca8\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509572 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27px\" (UniqueName: \"kubernetes.io/projected/fd146b1e-59a9-4246-9520-f2d6f6cf6cd1-kube-api-access-g27px\") pod \"ironic-operator-controller-manager-6f787dddc9-pcscc\" (UID: \"fd146b1e-59a9-4246-9520-f2d6f6cf6cd1\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509598 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv7tf\" (UniqueName: \"kubernetes.io/projected/ae8a8fbc-d425-4da5-afb3-438a85a43722-kube-api-access-dv7tf\") pod \"manila-operator-controller-manager-55f864c847-chmbd\" (UID: \"ae8a8fbc-d425-4da5-afb3-438a85a43722\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509618 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4sjl\" (UniqueName: \"kubernetes.io/projected/838f8a80-01c0-41d8-b431-2a23c9235fab-kube-api-access-w4sjl\") pod \"horizon-operator-controller-manager-8464cc45fb-wkjtf\" (UID: \"838f8a80-01c0-41d8-b431-2a23c9235fab\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.509654 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8pd6\" (UniqueName: \"kubernetes.io/projected/de7565f5-677b-4aeb-90ab-0d632b28b295-kube-api-access-p8pd6\") pod \"keystone-operator-controller-manager-768b96df4c-tp4s7\" (UID: \"de7565f5-677b-4aeb-90ab-0d632b28b295\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.524311 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.542851 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-gkndg"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.543771 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.545120 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.547440 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jbq96" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.571738 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.572713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.579400 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8kpfz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.580161 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.590110 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-gkndg"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.609298 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611644 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2plph\" (UniqueName: \"kubernetes.io/projected/bd5ae902-d036-4e52-983d-aa3e1a86dca8-kube-api-access-2plph\") pod \"heat-operator-controller-manager-67dd5f86f5-cxlns\" (UID: \"bd5ae902-d036-4e52-983d-aa3e1a86dca8\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27px\" (UniqueName: \"kubernetes.io/projected/fd146b1e-59a9-4246-9520-f2d6f6cf6cd1-kube-api-access-g27px\") pod \"ironic-operator-controller-manager-6f787dddc9-pcscc\" (UID: \"fd146b1e-59a9-4246-9520-f2d6f6cf6cd1\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611727 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcwp7\" (UniqueName: \"kubernetes.io/projected/f93025ae-ebc3-4aed-bfde-e514d8b814ce-kube-api-access-lcwp7\") pod \"nova-operator-controller-manager-5d488d59fb-jmwdk\" (UID: \"f93025ae-ebc3-4aed-bfde-e514d8b814ce\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611769 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv7tf\" (UniqueName: \"kubernetes.io/projected/ae8a8fbc-d425-4da5-afb3-438a85a43722-kube-api-access-dv7tf\") pod \"manila-operator-controller-manager-55f864c847-chmbd\" (UID: \"ae8a8fbc-d425-4da5-afb3-438a85a43722\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611796 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4sjl\" (UniqueName: \"kubernetes.io/projected/838f8a80-01c0-41d8-b431-2a23c9235fab-kube-api-access-w4sjl\") pod \"horizon-operator-controller-manager-8464cc45fb-wkjtf\" (UID: \"838f8a80-01c0-41d8-b431-2a23c9235fab\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8pd6\" (UniqueName: \"kubernetes.io/projected/de7565f5-677b-4aeb-90ab-0d632b28b295-kube-api-access-p8pd6\") pod \"keystone-operator-controller-manager-768b96df4c-tp4s7\" (UID: \"de7565f5-677b-4aeb-90ab-0d632b28b295\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqvtv\" (UniqueName: \"kubernetes.io/projected/216f9239-7d2e-483e-a89f-0955a518aa4a-kube-api-access-qqvtv\") pod \"neutron-operator-controller-manager-767865f676-gkndg\" (UID: \"216f9239-7d2e-483e-a89f-0955a518aa4a\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktf2p\" (UniqueName: \"kubernetes.io/projected/79dfdcde-0538-4777-959e-1daf2b6263de-kube-api-access-ktf2p\") pod \"mariadb-operator-controller-manager-67ccfc9778-4xzlc\" (UID: \"79dfdcde-0538-4777-959e-1daf2b6263de\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.611979 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnppd\" (UniqueName: \"kubernetes.io/projected/651c7dd5-3adc-48b4-b579-309258aa3735-kube-api-access-pnppd\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: E0318 10:29:36.612888 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:36 crc kubenswrapper[4733]: E0318 10:29:36.612940 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:37.112921979 +0000 UTC m=+1016.604656304 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.624772 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.625750 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.629991 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.635040 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-ft7lx" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.640446 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv7tf\" (UniqueName: \"kubernetes.io/projected/ae8a8fbc-d425-4da5-afb3-438a85a43722-kube-api-access-dv7tf\") pod \"manila-operator-controller-manager-55f864c847-chmbd\" (UID: \"ae8a8fbc-d425-4da5-afb3-438a85a43722\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.643578 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8pd6\" (UniqueName: \"kubernetes.io/projected/de7565f5-677b-4aeb-90ab-0d632b28b295-kube-api-access-p8pd6\") pod \"keystone-operator-controller-manager-768b96df4c-tp4s7\" (UID: \"de7565f5-677b-4aeb-90ab-0d632b28b295\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.646359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktf2p\" (UniqueName: \"kubernetes.io/projected/79dfdcde-0538-4777-959e-1daf2b6263de-kube-api-access-ktf2p\") pod \"mariadb-operator-controller-manager-67ccfc9778-4xzlc\" (UID: \"79dfdcde-0538-4777-959e-1daf2b6263de\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.648116 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.649224 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27px\" (UniqueName: \"kubernetes.io/projected/fd146b1e-59a9-4246-9520-f2d6f6cf6cd1-kube-api-access-g27px\") pod \"ironic-operator-controller-manager-6f787dddc9-pcscc\" (UID: \"fd146b1e-59a9-4246-9520-f2d6f6cf6cd1\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.648749 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2plph\" (UniqueName: \"kubernetes.io/projected/bd5ae902-d036-4e52-983d-aa3e1a86dca8-kube-api-access-2plph\") pod \"heat-operator-controller-manager-67dd5f86f5-cxlns\" (UID: \"bd5ae902-d036-4e52-983d-aa3e1a86dca8\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.649663 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4sjl\" (UniqueName: \"kubernetes.io/projected/838f8a80-01c0-41d8-b431-2a23c9235fab-kube-api-access-w4sjl\") pod \"horizon-operator-controller-manager-8464cc45fb-wkjtf\" (UID: \"838f8a80-01c0-41d8-b431-2a23c9235fab\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.655938 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.661064 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-gwtrb" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.668019 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-flv24"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.673452 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.675563 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.675933 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnppd\" (UniqueName: \"kubernetes.io/projected/651c7dd5-3adc-48b4-b579-309258aa3735-kube-api-access-pnppd\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.679946 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vdfsq" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.683817 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.695669 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.702111 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-flv24"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.709366 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9txbj"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.710386 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.711097 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712511 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqvtv\" (UniqueName: \"kubernetes.io/projected/216f9239-7d2e-483e-a89f-0955a518aa4a-kube-api-access-qqvtv\") pod \"neutron-operator-controller-manager-767865f676-gkndg\" (UID: \"216f9239-7d2e-483e-a89f-0955a518aa4a\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712570 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcwp7\" (UniqueName: \"kubernetes.io/projected/f93025ae-ebc3-4aed-bfde-e514d8b814ce-kube-api-access-lcwp7\") pod \"nova-operator-controller-manager-5d488d59fb-jmwdk\" (UID: \"f93025ae-ebc3-4aed-bfde-e514d8b814ce\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54rxf\" (UniqueName: \"kubernetes.io/projected/4ad2d88a-c733-4409-b07b-5ff4661e1b68-kube-api-access-54rxf\") pod \"placement-operator-controller-manager-5784578c99-9txbj\" (UID: \"4ad2d88a-c733-4409-b07b-5ff4661e1b68\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712633 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc2l9\" (UniqueName: \"kubernetes.io/projected/31999dbe-554e-4168-a902-1f62e82ce854-kube-api-access-rc2l9\") pod \"octavia-operator-controller-manager-5b9f45d989-22wt5\" (UID: \"31999dbe-554e-4168-a902-1f62e82ce854\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712667 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntcl4\" (UniqueName: \"kubernetes.io/projected/6762c515-b422-4157-a8ce-b9ca4781e134-kube-api-access-ntcl4\") pod \"ovn-operator-controller-manager-884679f54-flv24\" (UID: \"6762c515-b422-4157-a8ce-b9ca4781e134\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712694 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h9v6\" (UniqueName: \"kubernetes.io/projected/6eca2f16-53b8-4173-ace4-18b7292b1369-kube-api-access-4h9v6\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.712762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-l6bcs" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.747254 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.757706 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9txbj"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.767036 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcwp7\" (UniqueName: \"kubernetes.io/projected/f93025ae-ebc3-4aed-bfde-e514d8b814ce-kube-api-access-lcwp7\") pod \"nova-operator-controller-manager-5d488d59fb-jmwdk\" (UID: \"f93025ae-ebc3-4aed-bfde-e514d8b814ce\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.781825 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.784786 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.785423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqvtv\" (UniqueName: \"kubernetes.io/projected/216f9239-7d2e-483e-a89f-0955a518aa4a-kube-api-access-qqvtv\") pod \"neutron-operator-controller-manager-767865f676-gkndg\" (UID: \"216f9239-7d2e-483e-a89f-0955a518aa4a\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.786345 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-qc8z2" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.798059 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.798892 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.810706 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jnvkk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.813302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntcl4\" (UniqueName: \"kubernetes.io/projected/6762c515-b422-4157-a8ce-b9ca4781e134-kube-api-access-ntcl4\") pod \"ovn-operator-controller-manager-884679f54-flv24\" (UID: \"6762c515-b422-4157-a8ce-b9ca4781e134\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.813347 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h9v6\" (UniqueName: \"kubernetes.io/projected/6eca2f16-53b8-4173-ace4-18b7292b1369-kube-api-access-4h9v6\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.813418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54rxf\" (UniqueName: \"kubernetes.io/projected/4ad2d88a-c733-4409-b07b-5ff4661e1b68-kube-api-access-54rxf\") pod \"placement-operator-controller-manager-5784578c99-9txbj\" (UID: \"4ad2d88a-c733-4409-b07b-5ff4661e1b68\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.813453 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.813469 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc2l9\" (UniqueName: \"kubernetes.io/projected/31999dbe-554e-4168-a902-1f62e82ce854-kube-api-access-rc2l9\") pod \"octavia-operator-controller-manager-5b9f45d989-22wt5\" (UID: \"31999dbe-554e-4168-a902-1f62e82ce854\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:36 crc kubenswrapper[4733]: E0318 10:29:36.813952 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:36 crc kubenswrapper[4733]: E0318 10:29:36.814013 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:37.313994376 +0000 UTC m=+1016.805728701 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.817943 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.822646 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.824598 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.835369 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntcl4\" (UniqueName: \"kubernetes.io/projected/6762c515-b422-4157-a8ce-b9ca4781e134-kube-api-access-ntcl4\") pod \"ovn-operator-controller-manager-884679f54-flv24\" (UID: \"6762c515-b422-4157-a8ce-b9ca4781e134\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.842116 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54rxf\" (UniqueName: \"kubernetes.io/projected/4ad2d88a-c733-4409-b07b-5ff4661e1b68-kube-api-access-54rxf\") pod \"placement-operator-controller-manager-5784578c99-9txbj\" (UID: \"4ad2d88a-c733-4409-b07b-5ff4661e1b68\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.843326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h9v6\" (UniqueName: \"kubernetes.io/projected/6eca2f16-53b8-4173-ace4-18b7292b1369-kube-api-access-4h9v6\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.853552 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc2l9\" (UniqueName: \"kubernetes.io/projected/31999dbe-554e-4168-a902-1f62e82ce854-kube-api-access-rc2l9\") pod \"octavia-operator-controller-manager-5b9f45d989-22wt5\" (UID: \"31999dbe-554e-4168-a902-1f62e82ce854\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.868884 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.870629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.873291 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.880625 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.880844 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vfhsx" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.894716 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.915578 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.915922 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsj46\" (UniqueName: \"kubernetes.io/projected/759f85a1-4e24-4b61-879b-90801d648683-kube-api-access-jsj46\") pod \"telemetry-operator-controller-manager-d6b694c5-fd4t7\" (UID: \"759f85a1-4e24-4b61-879b-90801d648683\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.916007 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7l8w\" (UniqueName: \"kubernetes.io/projected/6ea742ac-3be9-4067-ab5a-032365494fde-kube-api-access-v7l8w\") pod \"swift-operator-controller-manager-c674c5965-v2pb2\" (UID: \"6ea742ac-3be9-4067-ab5a-032365494fde\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.917389 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.921303 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.922128 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.924518 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qhdl6" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.936443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.938582 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2"] Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.942026 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.948381 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.948557 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.948666 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5xhfs" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.955598 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:36 crc kubenswrapper[4733]: I0318 10:29:36.966255 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2"] Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.001698 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch"] Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.002559 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.007931 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6f5z6" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.008451 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch"] Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.017583 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7l8w\" (UniqueName: \"kubernetes.io/projected/6ea742ac-3be9-4067-ab5a-032365494fde-kube-api-access-v7l8w\") pod \"swift-operator-controller-manager-c674c5965-v2pb2\" (UID: \"6ea742ac-3be9-4067-ab5a-032365494fde\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.017638 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc79f\" (UniqueName: \"kubernetes.io/projected/6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54-kube-api-access-cc79f\") pod \"test-operator-controller-manager-5c5cb9c4d7-nskpj\" (UID: \"6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.017710 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsj46\" (UniqueName: \"kubernetes.io/projected/759f85a1-4e24-4b61-879b-90801d648683-kube-api-access-jsj46\") pod \"telemetry-operator-controller-manager-d6b694c5-fd4t7\" (UID: \"759f85a1-4e24-4b61-879b-90801d648683\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.061574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsj46\" (UniqueName: \"kubernetes.io/projected/759f85a1-4e24-4b61-879b-90801d648683-kube-api-access-jsj46\") pod \"telemetry-operator-controller-manager-d6b694c5-fd4t7\" (UID: \"759f85a1-4e24-4b61-879b-90801d648683\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.064399 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7l8w\" (UniqueName: \"kubernetes.io/projected/6ea742ac-3be9-4067-ab5a-032365494fde-kube-api-access-v7l8w\") pod \"swift-operator-controller-manager-c674c5965-v2pb2\" (UID: \"6ea742ac-3be9-4067-ab5a-032365494fde\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.084493 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.112515 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119521 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc79f\" (UniqueName: \"kubernetes.io/projected/6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54-kube-api-access-cc79f\") pod \"test-operator-controller-manager-5c5cb9c4d7-nskpj\" (UID: \"6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119679 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psssn\" (UniqueName: \"kubernetes.io/projected/cd9234ed-fcbc-4d81-9034-27d39b3df6ee-kube-api-access-psssn\") pod \"watcher-operator-controller-manager-6c4d75f7f9-sqr4g\" (UID: \"cd9234ed-fcbc-4d81-9034-27d39b3df6ee\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119733 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cb9b\" (UniqueName: \"kubernetes.io/projected/a4b7e706-a9a7-490a-84a8-094d1d909ba8-kube-api-access-4cb9b\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119766 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brp9l\" (UniqueName: \"kubernetes.io/projected/e64c7cd6-a04b-440e-ac47-40f672fbc333-kube-api-access-brp9l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-k64ch\" (UID: \"e64c7cd6-a04b-440e-ac47-40f672fbc333\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.119812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.119959 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.120010 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:38.119992109 +0000 UTC m=+1017.611726434 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.141220 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc79f\" (UniqueName: \"kubernetes.io/projected/6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54-kube-api-access-cc79f\") pod \"test-operator-controller-manager-5c5cb9c4d7-nskpj\" (UID: \"6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.170869 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.172736 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.222734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.222779 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psssn\" (UniqueName: \"kubernetes.io/projected/cd9234ed-fcbc-4d81-9034-27d39b3df6ee-kube-api-access-psssn\") pod \"watcher-operator-controller-manager-6c4d75f7f9-sqr4g\" (UID: \"cd9234ed-fcbc-4d81-9034-27d39b3df6ee\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.222797 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cb9b\" (UniqueName: \"kubernetes.io/projected/a4b7e706-a9a7-490a-84a8-094d1d909ba8-kube-api-access-4cb9b\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.222831 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brp9l\" (UniqueName: \"kubernetes.io/projected/e64c7cd6-a04b-440e-ac47-40f672fbc333-kube-api-access-brp9l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-k64ch\" (UID: \"e64c7cd6-a04b-440e-ac47-40f672fbc333\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.222866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.222928 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.223000 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:37.722983342 +0000 UTC m=+1017.214717667 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.223012 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.223055 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:37.723043083 +0000 UTC m=+1017.214777398 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.229439 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn"] Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.245231 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.249065 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psssn\" (UniqueName: \"kubernetes.io/projected/cd9234ed-fcbc-4d81-9034-27d39b3df6ee-kube-api-access-psssn\") pod \"watcher-operator-controller-manager-6c4d75f7f9-sqr4g\" (UID: \"cd9234ed-fcbc-4d81-9034-27d39b3df6ee\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.250011 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brp9l\" (UniqueName: \"kubernetes.io/projected/e64c7cd6-a04b-440e-ac47-40f672fbc333-kube-api-access-brp9l\") pod \"rabbitmq-cluster-operator-manager-668c99d594-k64ch\" (UID: \"e64c7cd6-a04b-440e-ac47-40f672fbc333\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.254370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cb9b\" (UniqueName: \"kubernetes.io/projected/a4b7e706-a9a7-490a-84a8-094d1d909ba8-kube-api-access-4cb9b\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.262144 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.325203 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.325377 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.325439 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:38.325419249 +0000 UTC m=+1017.817153574 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: W0318 10:29:37.326587 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fe910c4_798b_4381_a71d_697459f7f79a.slice/crio-3f68db198b7463ac4b05d69c1568aa5c5301de1dce7b45824d44840ffbb5e0c5 WatchSource:0}: Error finding container 3f68db198b7463ac4b05d69c1568aa5c5301de1dce7b45824d44840ffbb5e0c5: Status 404 returned error can't find the container with id 3f68db198b7463ac4b05d69c1568aa5c5301de1dce7b45824d44840ffbb5e0c5 Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.433847 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" event={"ID":"8fe910c4-798b-4381-a71d-697459f7f79a","Type":"ContainerStarted","Data":"3f68db198b7463ac4b05d69c1568aa5c5301de1dce7b45824d44840ffbb5e0c5"} Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.485616 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.734829 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.734901 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.735032 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.735112 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:38.735098655 +0000 UTC m=+1018.226832980 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.736796 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: E0318 10:29:37.736897 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:38.736877676 +0000 UTC m=+1018.228612001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.754763 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-t8796"] Mar 18 10:29:37 crc kubenswrapper[4733]: I0318 10:29:37.796671 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v"] Mar 18 10:29:37 crc kubenswrapper[4733]: W0318 10:29:37.800527 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fb2ba68_fa0f_4483_afdf_2eb381c54320.slice/crio-088842a868b02ec21f48d7509be6d13ba2f98f0fb318430fd498f2f31804837a WatchSource:0}: Error finding container 088842a868b02ec21f48d7509be6d13ba2f98f0fb318430fd498f2f31804837a: Status 404 returned error can't find the container with id 088842a868b02ec21f48d7509be6d13ba2f98f0fb318430fd498f2f31804837a Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.090949 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.099111 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.105611 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.127642 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.144019 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.144311 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.144359 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:40.14434358 +0000 UTC m=+1019.636077905 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.168781 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.187117 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.200850 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-gkndg"] Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.204178 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod216f9239_7d2e_483e_a89f_0955a518aa4a.slice/crio-717164cede3a5aeefb57345884e06771787e1902761118a64f9a630f0786dfab WatchSource:0}: Error finding container 717164cede3a5aeefb57345884e06771787e1902761118a64f9a630f0786dfab: Status 404 returned error can't find the container with id 717164cede3a5aeefb57345884e06771787e1902761118a64f9a630f0786dfab Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.204398 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf93025ae_ebc3_4aed_bfde_e514d8b814ce.slice/crio-7923bc481871d27755290fc6dca542bca37b41aefe9f9c9377630357efee6cc5 WatchSource:0}: Error finding container 7923bc481871d27755290fc6dca542bca37b41aefe9f9c9377630357efee6cc5: Status 404 returned error can't find the container with id 7923bc481871d27755290fc6dca542bca37b41aefe9f9c9377630357efee6cc5 Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.206044 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae8a8fbc_d425_4da5_afb3_438a85a43722.slice/crio-52924df880af4e3e594152ddc2ac8ec42109dc8f64facd9860786f838b963f8c WatchSource:0}: Error finding container 52924df880af4e3e594152ddc2ac8ec42109dc8f64facd9860786f838b963f8c: Status 404 returned error can't find the container with id 52924df880af4e3e594152ddc2ac8ec42109dc8f64facd9860786f838b963f8c Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.212296 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dv7tf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-55f864c847-chmbd_openstack-operators(ae8a8fbc-d425-4da5-afb3-438a85a43722): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.215048 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk"] Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.215126 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" podUID="ae8a8fbc-d425-4da5-afb3-438a85a43722" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.219833 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-chmbd"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.225032 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-flv24"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.228465 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.296096 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2"] Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.305549 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj"] Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.310352 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jsj46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-fd4t7_openstack-operators(759f85a1-4e24-4b61-879b-90801d648683): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.311789 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" podUID="759f85a1-4e24-4b61-879b-90801d648683" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.312274 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7"] Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.312799 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cc79f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-nskpj_openstack-operators(6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.314573 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podUID="6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54" Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.317613 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea742ac_3be9_4067_ab5a_032365494fde.slice/crio-a4a2daa0e2c90160ccc562b8b047470f9d8ba2da90c1e739c76cddd6f5c0d339 WatchSource:0}: Error finding container a4a2daa0e2c90160ccc562b8b047470f9d8ba2da90c1e739c76cddd6f5c0d339: Status 404 returned error can't find the container with id a4a2daa0e2c90160ccc562b8b047470f9d8ba2da90c1e739c76cddd6f5c0d339 Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.319993 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch"] Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.324689 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad2d88a_c733_4409_b07b_5ff4661e1b68.slice/crio-810ef674bbeea3daf9a479b03dc0f650895e68ece3f19c3fba39a64077d8a2da WatchSource:0}: Error finding container 810ef674bbeea3daf9a479b03dc0f650895e68ece3f19c3fba39a64077d8a2da: Status 404 returned error can't find the container with id 810ef674bbeea3daf9a479b03dc0f650895e68ece3f19c3fba39a64077d8a2da Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.327612 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-9txbj"] Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.329596 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brp9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-k64ch_openstack-operators(e64c7cd6-a04b-440e-ac47-40f672fbc333): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.329764 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-54rxf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-9txbj_openstack-operators(4ad2d88a-c733-4409-b07b-5ff4661e1b68): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.330547 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v7l8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-c674c5965-v2pb2_openstack-operators(6ea742ac-3be9-4067-ab5a-032365494fde): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.331612 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" podUID="4ad2d88a-c733-4409-b07b-5ff4661e1b68" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.331672 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" podUID="6ea742ac-3be9-4067-ab5a-032365494fde" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.331703 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" podUID="e64c7cd6-a04b-440e-ac47-40f672fbc333" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.333545 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g"] Mar 18 10:29:38 crc kubenswrapper[4733]: W0318 10:29:38.334025 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd9234ed_fcbc_4d81_9034_27d39b3df6ee.slice/crio-f653d2a2a5115dd8636cc652178e2a29bc712a9f89a4aa8314c3a96375f2887f WatchSource:0}: Error finding container f653d2a2a5115dd8636cc652178e2a29bc712a9f89a4aa8314c3a96375f2887f: Status 404 returned error can't find the container with id f653d2a2a5115dd8636cc652178e2a29bc712a9f89a4aa8314c3a96375f2887f Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.336012 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-psssn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-sqr4g_openstack-operators(cd9234ed-fcbc-4d81-9034-27d39b3df6ee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.337197 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podUID="cd9234ed-fcbc-4d81-9034-27d39b3df6ee" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.360316 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.360505 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.360557 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:40.360540975 +0000 UTC m=+1019.852275310 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.443783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" event={"ID":"ae8a8fbc-d425-4da5-afb3-438a85a43722","Type":"ContainerStarted","Data":"52924df880af4e3e594152ddc2ac8ec42109dc8f64facd9860786f838b963f8c"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.446434 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" podUID="ae8a8fbc-d425-4da5-afb3-438a85a43722" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.448925 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" event={"ID":"6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54","Type":"ContainerStarted","Data":"b1c5ce101e4b02d694f4e46490df1e2a8d3c161b0cea82fab9664954d93969c4"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.450657 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podUID="6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.451848 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" event={"ID":"759f85a1-4e24-4b61-879b-90801d648683","Type":"ContainerStarted","Data":"99b01cc5f636ca11bb07336e824f83b58a31e7c5eb67b8193ba81955c51fdacc"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.453690 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" event={"ID":"216f9239-7d2e-483e-a89f-0955a518aa4a","Type":"ContainerStarted","Data":"717164cede3a5aeefb57345884e06771787e1902761118a64f9a630f0786dfab"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.454384 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" podUID="759f85a1-4e24-4b61-879b-90801d648683" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.454960 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" event={"ID":"6ea742ac-3be9-4067-ab5a-032365494fde","Type":"ContainerStarted","Data":"a4a2daa0e2c90160ccc562b8b047470f9d8ba2da90c1e739c76cddd6f5c0d339"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.455873 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" event={"ID":"fd146b1e-59a9-4246-9520-f2d6f6cf6cd1","Type":"ContainerStarted","Data":"42b2fa3897738c27aae27e64fe8fd12e27930763f1c5ab42d550a86a108f7054"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.456245 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" podUID="6ea742ac-3be9-4067-ab5a-032365494fde" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.461248 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" event={"ID":"4ad2d88a-c733-4409-b07b-5ff4661e1b68","Type":"ContainerStarted","Data":"810ef674bbeea3daf9a479b03dc0f650895e68ece3f19c3fba39a64077d8a2da"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.463041 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" podUID="4ad2d88a-c733-4409-b07b-5ff4661e1b68" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.463732 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" event={"ID":"de7565f5-677b-4aeb-90ab-0d632b28b295","Type":"ContainerStarted","Data":"7c8b4e22673abb3f2244e1c43a8b76bdb4ef12d3caa4a97558cbbf7d2372e1f7"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.466460 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" event={"ID":"f93025ae-ebc3-4aed-bfde-e514d8b814ce","Type":"ContainerStarted","Data":"7923bc481871d27755290fc6dca542bca37b41aefe9f9c9377630357efee6cc5"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.468030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" event={"ID":"0fb2ba68-fa0f-4483-afdf-2eb381c54320","Type":"ContainerStarted","Data":"088842a868b02ec21f48d7509be6d13ba2f98f0fb318430fd498f2f31804837a"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.472180 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" event={"ID":"bc0e28fc-cff0-4c39-8073-61d5d6481866","Type":"ContainerStarted","Data":"0dcd0d8305b68c607bd9507592f538c00ad22c5eee49e0dc66436e89416262f9"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.476553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" event={"ID":"748f4855-3978-4ecd-805e-0fee34ce0094","Type":"ContainerStarted","Data":"d0174c0fd3f0a67e352fae83a5dcec557c52831c71be828615bf69e0cd5d5c94"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.478118 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" event={"ID":"838f8a80-01c0-41d8-b431-2a23c9235fab","Type":"ContainerStarted","Data":"26393f7f6af834af39a55ae9cf8af4ddf7cc7d8a2b57dfc8c2a817b2a41425bd"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.487270 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" event={"ID":"6762c515-b422-4157-a8ce-b9ca4781e134","Type":"ContainerStarted","Data":"f18e7a9763b0f4e2f719e251cdfefbaab95d7906fe1753a5def0e9d13fcf2f00"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.489908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" event={"ID":"31999dbe-554e-4168-a902-1f62e82ce854","Type":"ContainerStarted","Data":"5206dc219d51946d37b6adb4bfce1ad2ae9878b57407a33030422fe6a224973c"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.492784 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" event={"ID":"e64c7cd6-a04b-440e-ac47-40f672fbc333","Type":"ContainerStarted","Data":"8e41e541f602b6562d26e9625b6313839135de78bad4176fe60420dda46a9b37"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.493933 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" podUID="e64c7cd6-a04b-440e-ac47-40f672fbc333" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.495263 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" event={"ID":"cd9234ed-fcbc-4d81-9034-27d39b3df6ee","Type":"ContainerStarted","Data":"f653d2a2a5115dd8636cc652178e2a29bc712a9f89a4aa8314c3a96375f2887f"} Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.498169 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podUID="cd9234ed-fcbc-4d81-9034-27d39b3df6ee" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.503808 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" event={"ID":"bd5ae902-d036-4e52-983d-aa3e1a86dca8","Type":"ContainerStarted","Data":"be9db7d7665259b9c7f9456c3fa4100b6ed161f66799a501bf6e6d7163bd9be4"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.507098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" event={"ID":"79dfdcde-0538-4777-959e-1daf2b6263de","Type":"ContainerStarted","Data":"fb165269bb33745d1b2a430bb177e8169a32c140810add4c6ba806589792fde4"} Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.770354 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:38 crc kubenswrapper[4733]: I0318 10:29:38.770455 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.770605 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.770709 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:40.770682036 +0000 UTC m=+1020.262416361 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.770707 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:38 crc kubenswrapper[4733]: E0318 10:29:38.770799 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:40.770774158 +0000 UTC m=+1020.262508513 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.519244 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" podUID="e64c7cd6-a04b-440e-ac47-40f672fbc333" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.519716 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podUID="cd9234ed-fcbc-4d81-9034-27d39b3df6ee" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.519741 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" podUID="759f85a1-4e24-4b61-879b-90801d648683" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.519763 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" podUID="6ea742ac-3be9-4067-ab5a-032365494fde" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.521600 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podUID="6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.521594 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" podUID="ae8a8fbc-d425-4da5-afb3-438a85a43722" Mar 18 10:29:39 crc kubenswrapper[4733]: E0318 10:29:39.547628 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" podUID="4ad2d88a-c733-4409-b07b-5ff4661e1b68" Mar 18 10:29:40 crc kubenswrapper[4733]: I0318 10:29:40.194973 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.195163 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.195252 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:44.195232674 +0000 UTC m=+1023.686966999 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: I0318 10:29:40.397923 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.398097 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.398166 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:44.398149803 +0000 UTC m=+1023.889884128 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: I0318 10:29:40.814680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:40 crc kubenswrapper[4733]: I0318 10:29:40.814763 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.814849 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.814910 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.814916 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:44.814900229 +0000 UTC m=+1024.306634554 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:40 crc kubenswrapper[4733]: E0318 10:29:40.814969 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:44.814955801 +0000 UTC m=+1024.306690126 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:43 crc kubenswrapper[4733]: I0318 10:29:43.570885 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:29:43 crc kubenswrapper[4733]: I0318 10:29:43.571323 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:29:44 crc kubenswrapper[4733]: I0318 10:29:44.268304 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.268528 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.268614 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:52.268590943 +0000 UTC m=+1031.760325338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: I0318 10:29:44.471589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.471839 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.472020 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:52.471997016 +0000 UTC m=+1031.963731361 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: I0318 10:29:44.876420 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:44 crc kubenswrapper[4733]: I0318 10:29:44.876587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.876635 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.876725 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:52.876702051 +0000 UTC m=+1032.368436446 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.876786 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:44 crc kubenswrapper[4733]: E0318 10:29:44.876854 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:29:52.876833655 +0000 UTC m=+1032.368568040 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.622637 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" event={"ID":"f93025ae-ebc3-4aed-bfde-e514d8b814ce","Type":"ContainerStarted","Data":"471ea8594a79d8f2f603499be05353299288503e28f7302da27e714c8e91c9f3"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.623163 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.633779 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" event={"ID":"31999dbe-554e-4168-a902-1f62e82ce854","Type":"ContainerStarted","Data":"9855c34d85194f685066208c952a3c912a12cac69f6aa9c2e16fa29a45dc6639"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.633918 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.636337 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" event={"ID":"0fb2ba68-fa0f-4483-afdf-2eb381c54320","Type":"ContainerStarted","Data":"bd08a464f25b997614e25816121dcdf634e8a1cdc018e7a9099a3dd0fcc112c1"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.636802 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.643514 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" event={"ID":"79dfdcde-0538-4777-959e-1daf2b6263de","Type":"ContainerStarted","Data":"c71bed7fb0c0f1badd4e2e99f5fdb45c98e995f1b2e899eea883d9051044b4d5"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.644067 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.648170 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" event={"ID":"bc0e28fc-cff0-4c39-8073-61d5d6481866","Type":"ContainerStarted","Data":"52486ba6a080e1684f6bf49c4cc34a4c30accc27c2f37344a6e7db1fb04a9ff2"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.648858 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.695070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" event={"ID":"de7565f5-677b-4aeb-90ab-0d632b28b295","Type":"ContainerStarted","Data":"73bbdd7a5a2137122c760ed8676dff0b80127d8b5f32cd1ed3d915ab84e65207"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.695822 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.700093 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" event={"ID":"bd5ae902-d036-4e52-983d-aa3e1a86dca8","Type":"ContainerStarted","Data":"9275c970ce11d7157c2d8eb0d744d0b6dbffe6ca0eb408e5963bb46c740847c2"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.700626 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.715440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" event={"ID":"8fe910c4-798b-4381-a71d-697459f7f79a","Type":"ContainerStarted","Data":"738fe7648e2881bb14f06a1f26c69cafb94899de265e9a0d967a8c1fc7512891"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.716134 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.721028 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" event={"ID":"216f9239-7d2e-483e-a89f-0955a518aa4a","Type":"ContainerStarted","Data":"e2fc3c58d162844f3cb91d62cbe3c85e1f5f2ef70322bba2afe7c5f97c10d1d5"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.721739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.723805 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" event={"ID":"fd146b1e-59a9-4246-9520-f2d6f6cf6cd1","Type":"ContainerStarted","Data":"f17da01172944e9d80d11af12daace16ec5a3506eaaaefab4432d565d3f6802d"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.724211 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.727396 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" event={"ID":"6762c515-b422-4157-a8ce-b9ca4781e134","Type":"ContainerStarted","Data":"0c36a7a036f48577f6a51a7367515826c8a5b92f44acf10407371c9c3f26f282"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.727819 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.727935 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" podStartSLOduration=3.168346439 podStartE2EDuration="15.727919735s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.209049506 +0000 UTC m=+1017.700783831" lastFinishedPulling="2026-03-18 10:29:50.768622782 +0000 UTC m=+1030.260357127" observedRunningTime="2026-03-18 10:29:51.726589307 +0000 UTC m=+1031.218323632" watchObservedRunningTime="2026-03-18 10:29:51.727919735 +0000 UTC m=+1031.219654060" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.754208 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" event={"ID":"748f4855-3978-4ecd-805e-0fee34ce0094","Type":"ContainerStarted","Data":"249205d73477e093aa2dbec75fa5a6f569b11b1f3effb2950ecd8e672a994f59"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.754254 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.765391 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" event={"ID":"838f8a80-01c0-41d8-b431-2a23c9235fab","Type":"ContainerStarted","Data":"68c85c5e0e8adc64073999b3dc1858a7182233bef9805a6deb0937d07a7a72df"} Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.766004 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:51 crc kubenswrapper[4733]: I0318 10:29:51.972032 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" podStartSLOduration=3.379777279 podStartE2EDuration="15.972012792s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.127306456 +0000 UTC m=+1017.619040781" lastFinishedPulling="2026-03-18 10:29:50.719541979 +0000 UTC m=+1030.211276294" observedRunningTime="2026-03-18 10:29:51.97123657 +0000 UTC m=+1031.462970895" watchObservedRunningTime="2026-03-18 10:29:51.972012792 +0000 UTC m=+1031.463747127" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.145839 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" podStartSLOduration=3.51262107 podStartE2EDuration="16.145816615s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.120787181 +0000 UTC m=+1017.612521496" lastFinishedPulling="2026-03-18 10:29:50.753982716 +0000 UTC m=+1030.245717041" observedRunningTime="2026-03-18 10:29:52.134864394 +0000 UTC m=+1031.626598719" watchObservedRunningTime="2026-03-18 10:29:52.145816615 +0000 UTC m=+1031.637550960" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.285124 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" podStartSLOduration=3.714139679 podStartE2EDuration="16.285106748s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.138599986 +0000 UTC m=+1017.630334311" lastFinishedPulling="2026-03-18 10:29:50.709567055 +0000 UTC m=+1030.201301380" observedRunningTime="2026-03-18 10:29:52.264513024 +0000 UTC m=+1031.756247349" watchObservedRunningTime="2026-03-18 10:29:52.285106748 +0000 UTC m=+1031.776841073" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.294101 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.294332 4733 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.294380 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert podName:651c7dd5-3adc-48b4-b579-309258aa3735 nodeName:}" failed. No retries permitted until 2026-03-18 10:30:08.294367061 +0000 UTC m=+1047.786101376 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert") pod "infra-operator-controller-manager-74c694b97b-j4snz" (UID: "651c7dd5-3adc-48b4-b579-309258aa3735") : secret "infra-operator-webhook-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.339458 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" podStartSLOduration=3.724522144 podStartE2EDuration="16.33943963s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.122120819 +0000 UTC m=+1017.613855144" lastFinishedPulling="2026-03-18 10:29:50.737038295 +0000 UTC m=+1030.228772630" observedRunningTime="2026-03-18 10:29:52.338523944 +0000 UTC m=+1031.830258269" watchObservedRunningTime="2026-03-18 10:29:52.33943963 +0000 UTC m=+1031.831173955" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.419834 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" podStartSLOduration=3.929982484 podStartE2EDuration="16.419820801s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.209038265 +0000 UTC m=+1017.700772590" lastFinishedPulling="2026-03-18 10:29:50.698876582 +0000 UTC m=+1030.190610907" observedRunningTime="2026-03-18 10:29:52.417534386 +0000 UTC m=+1031.909268711" watchObservedRunningTime="2026-03-18 10:29:52.419820801 +0000 UTC m=+1031.911555126" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.421624 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" podStartSLOduration=3.830720137 podStartE2EDuration="16.421619662s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.119140414 +0000 UTC m=+1017.610874739" lastFinishedPulling="2026-03-18 10:29:50.710039939 +0000 UTC m=+1030.201774264" observedRunningTime="2026-03-18 10:29:52.372087587 +0000 UTC m=+1031.863821912" watchObservedRunningTime="2026-03-18 10:29:52.421619662 +0000 UTC m=+1031.913353977" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.441983 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" podStartSLOduration=3.711302458 podStartE2EDuration="16.44196827s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:37.362369997 +0000 UTC m=+1016.854104322" lastFinishedPulling="2026-03-18 10:29:50.093035809 +0000 UTC m=+1029.584770134" observedRunningTime="2026-03-18 10:29:52.439251473 +0000 UTC m=+1031.930985798" watchObservedRunningTime="2026-03-18 10:29:52.44196827 +0000 UTC m=+1031.933702595" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.466043 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" podStartSLOduration=3.978026058 podStartE2EDuration="16.465975531s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.198865527 +0000 UTC m=+1017.690599852" lastFinishedPulling="2026-03-18 10:29:50.686815 +0000 UTC m=+1030.178549325" observedRunningTime="2026-03-18 10:29:52.462966866 +0000 UTC m=+1031.954701201" watchObservedRunningTime="2026-03-18 10:29:52.465975531 +0000 UTC m=+1031.957709856" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.496299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.496438 4733 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.496495 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert podName:6eca2f16-53b8-4173-ace4-18b7292b1369 nodeName:}" failed. No retries permitted until 2026-03-18 10:30:08.496479307 +0000 UTC m=+1047.988213622 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" (UID: "6eca2f16-53b8-4173-ace4-18b7292b1369") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.504290 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" podStartSLOduration=4.213550232 podStartE2EDuration="16.504274258s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:37.802366394 +0000 UTC m=+1017.294100719" lastFinishedPulling="2026-03-18 10:29:50.09309042 +0000 UTC m=+1029.584824745" observedRunningTime="2026-03-18 10:29:52.488933012 +0000 UTC m=+1031.980667337" watchObservedRunningTime="2026-03-18 10:29:52.504274258 +0000 UTC m=+1031.996008573" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.509472 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" podStartSLOduration=3.926944169 podStartE2EDuration="16.509460435s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.155689412 +0000 UTC m=+1017.647423727" lastFinishedPulling="2026-03-18 10:29:50.738205668 +0000 UTC m=+1030.229939993" observedRunningTime="2026-03-18 10:29:52.504089563 +0000 UTC m=+1031.995823888" watchObservedRunningTime="2026-03-18 10:29:52.509460435 +0000 UTC m=+1032.001194750" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.544803 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" podStartSLOduration=4.040355166 podStartE2EDuration="16.544783747s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.20426993 +0000 UTC m=+1017.696004255" lastFinishedPulling="2026-03-18 10:29:50.708698521 +0000 UTC m=+1030.200432836" observedRunningTime="2026-03-18 10:29:52.544604182 +0000 UTC m=+1032.036338507" watchObservedRunningTime="2026-03-18 10:29:52.544783747 +0000 UTC m=+1032.036518072" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.564134 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" podStartSLOduration=3.581592067 podStartE2EDuration="16.564117836s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:37.765015734 +0000 UTC m=+1017.256750059" lastFinishedPulling="2026-03-18 10:29:50.747541503 +0000 UTC m=+1030.239275828" observedRunningTime="2026-03-18 10:29:52.559417333 +0000 UTC m=+1032.051151658" watchObservedRunningTime="2026-03-18 10:29:52.564117836 +0000 UTC m=+1032.055852161" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.906042 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:52 crc kubenswrapper[4733]: I0318 10:29:52.906288 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.907375 4733 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.907435 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:30:08.907419219 +0000 UTC m=+1048.399153544 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "metrics-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.907780 4733 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 10:29:52 crc kubenswrapper[4733]: E0318 10:29:52.907847 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs podName:a4b7e706-a9a7-490a-84a8-094d1d909ba8 nodeName:}" failed. No retries permitted until 2026-03-18 10:30:08.907830131 +0000 UTC m=+1048.399564456 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs") pod "openstack-operator-controller-manager-85877db48-qvlf2" (UID: "a4b7e706-a9a7-490a-84a8-094d1d909ba8") : secret "webhook-server-cert" not found Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.498682 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-sfv8v" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.505411 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-v6zxn" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.550912 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-t8796" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.584753 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-ljvrt" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.699750 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-cxlns" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.716349 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-wkjtf" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.750229 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-pcscc" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.875341 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-tp4s7" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.898393 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4xzlc" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.920854 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.941476 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-jmwdk" Mar 18 10:29:56 crc kubenswrapper[4733]: I0318 10:29:56.964483 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-22wt5" Mar 18 10:29:57 crc kubenswrapper[4733]: I0318 10:29:57.096915 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-flv24" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.141471 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563830-2qqd2"] Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.142835 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.145106 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.145422 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.145589 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.155502 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563830-2qqd2"] Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.164362 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh"] Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.165133 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.167019 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.167693 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.183717 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh"] Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.221457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.221538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz2gq\" (UniqueName: \"kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq\") pod \"auto-csr-approver-29563830-2qqd2\" (UID: \"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35\") " pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.221607 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.221669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz865\" (UniqueName: \"kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.323095 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz2gq\" (UniqueName: \"kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq\") pod \"auto-csr-approver-29563830-2qqd2\" (UID: \"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35\") " pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.323237 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.323315 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz865\" (UniqueName: \"kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.323363 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.324569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.339755 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.341874 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz2gq\" (UniqueName: \"kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq\") pod \"auto-csr-approver-29563830-2qqd2\" (UID: \"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35\") " pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.346531 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz865\" (UniqueName: \"kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865\") pod \"collect-profiles-29563830-kdgdh\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.523075 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.531859 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.838070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" event={"ID":"759f85a1-4e24-4b61-879b-90801d648683","Type":"ContainerStarted","Data":"524add925102e84399badc3832c397a565f5c8e3d285558bd513f60adefbe52c"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.838471 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.839523 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" event={"ID":"6ea742ac-3be9-4067-ab5a-032365494fde","Type":"ContainerStarted","Data":"09cefc9491e9ff975d4e45fc5e9f851f90830f8364ec6c4df8421db2f20d0b62"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.839718 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.841163 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" event={"ID":"ae8a8fbc-d425-4da5-afb3-438a85a43722","Type":"ContainerStarted","Data":"e63d7e02138093bf3a90cb47a652968ea8c4f8c14758e128a67af5acc0b36599"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.841393 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.864723 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" podStartSLOduration=3.55700833 podStartE2EDuration="24.864706621s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.310227847 +0000 UTC m=+1017.801962172" lastFinishedPulling="2026-03-18 10:29:59.617926138 +0000 UTC m=+1039.109660463" observedRunningTime="2026-03-18 10:30:00.860603255 +0000 UTC m=+1040.352337580" watchObservedRunningTime="2026-03-18 10:30:00.864706621 +0000 UTC m=+1040.356440946" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.865930 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" event={"ID":"4ad2d88a-c733-4409-b07b-5ff4661e1b68","Type":"ContainerStarted","Data":"0337ff8f516557269758798f022de29676d9583baad256014db066578e9e029c"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.866500 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.869481 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" event={"ID":"e64c7cd6-a04b-440e-ac47-40f672fbc333","Type":"ContainerStarted","Data":"773d54dfac2ac55d8e5d7689898349f0d7536703793e842660114b5e23d74f56"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.872095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" event={"ID":"cd9234ed-fcbc-4d81-9034-27d39b3df6ee","Type":"ContainerStarted","Data":"70b824e8acd9e58d6693782cc83587264fbb5786b581efbfd2c997d24705f6f5"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.872454 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.873356 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" event={"ID":"6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54","Type":"ContainerStarted","Data":"2a4b3510fbc76543281dfd4f9a396cd04735643b77c8b669c851732e62a021d9"} Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.873666 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.894994 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" podStartSLOduration=3.518320477 podStartE2EDuration="24.894976837s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.212130203 +0000 UTC m=+1017.703864528" lastFinishedPulling="2026-03-18 10:29:59.588786563 +0000 UTC m=+1039.080520888" observedRunningTime="2026-03-18 10:30:00.889096681 +0000 UTC m=+1040.380831006" watchObservedRunningTime="2026-03-18 10:30:00.894976837 +0000 UTC m=+1040.386711162" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.919935 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" podStartSLOduration=3.700862483 podStartE2EDuration="24.919919643s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.330328998 +0000 UTC m=+1017.822063323" lastFinishedPulling="2026-03-18 10:29:59.549386118 +0000 UTC m=+1039.041120483" observedRunningTime="2026-03-18 10:30:00.91943954 +0000 UTC m=+1040.411173855" watchObservedRunningTime="2026-03-18 10:30:00.919919643 +0000 UTC m=+1040.411653968" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.958471 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" podStartSLOduration=3.743723086 podStartE2EDuration="24.958448804s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.329673179 +0000 UTC m=+1017.821407504" lastFinishedPulling="2026-03-18 10:29:59.544398897 +0000 UTC m=+1039.036133222" observedRunningTime="2026-03-18 10:30:00.945781885 +0000 UTC m=+1040.437516210" watchObservedRunningTime="2026-03-18 10:30:00.958448804 +0000 UTC m=+1040.450183129" Mar 18 10:30:00 crc kubenswrapper[4733]: I0318 10:30:00.982055 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podStartSLOduration=3.656803724 podStartE2EDuration="24.982030591s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.312693597 +0000 UTC m=+1017.804427912" lastFinishedPulling="2026-03-18 10:29:59.637920454 +0000 UTC m=+1039.129654779" observedRunningTime="2026-03-18 10:30:00.97846585 +0000 UTC m=+1040.470200175" watchObservedRunningTime="2026-03-18 10:30:00.982030591 +0000 UTC m=+1040.473764916" Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.012613 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-k64ch" podStartSLOduration=3.7992002449999998 podStartE2EDuration="25.012597806s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.329434382 +0000 UTC m=+1017.821168707" lastFinishedPulling="2026-03-18 10:29:59.542831933 +0000 UTC m=+1039.034566268" observedRunningTime="2026-03-18 10:30:01.010548938 +0000 UTC m=+1040.502283263" watchObservedRunningTime="2026-03-18 10:30:01.012597806 +0000 UTC m=+1040.504332121" Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.055704 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563830-2qqd2"] Mar 18 10:30:01 crc kubenswrapper[4733]: W0318 10:30:01.060861 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c8eb139_c576_4ab3_8c2b_a309f3aa4a35.slice/crio-8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765 WatchSource:0}: Error finding container 8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765: Status 404 returned error can't find the container with id 8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765 Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.091907 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podStartSLOduration=3.844421686 podStartE2EDuration="25.09189305s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:29:38.335897806 +0000 UTC m=+1017.827632131" lastFinishedPulling="2026-03-18 10:29:59.58336917 +0000 UTC m=+1039.075103495" observedRunningTime="2026-03-18 10:30:01.073356115 +0000 UTC m=+1040.565090440" watchObservedRunningTime="2026-03-18 10:30:01.09189305 +0000 UTC m=+1040.583627375" Mar 18 10:30:01 crc kubenswrapper[4733]: W0318 10:30:01.098854 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a150895_3ed5_4f19_8a97_3b65f3254672.slice/crio-1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440 WatchSource:0}: Error finding container 1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440: Status 404 returned error can't find the container with id 1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440 Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.099263 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh"] Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.883106 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" event={"ID":"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35","Type":"ContainerStarted","Data":"8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765"} Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.885055 4733 generic.go:334] "Generic (PLEG): container finished" podID="4a150895-3ed5-4f19-8a97-3b65f3254672" containerID="1afbfc2e0fbf329edca7836eaa8d44fdbd1ae521356c2969dfea34427ac61545" exitCode=0 Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.885172 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" event={"ID":"4a150895-3ed5-4f19-8a97-3b65f3254672","Type":"ContainerDied","Data":"1afbfc2e0fbf329edca7836eaa8d44fdbd1ae521356c2969dfea34427ac61545"} Mar 18 10:30:01 crc kubenswrapper[4733]: I0318 10:30:01.885310 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" event={"ID":"4a150895-3ed5-4f19-8a97-3b65f3254672","Type":"ContainerStarted","Data":"1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440"} Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.154485 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.271013 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz865\" (UniqueName: \"kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865\") pod \"4a150895-3ed5-4f19-8a97-3b65f3254672\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.271588 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume\") pod \"4a150895-3ed5-4f19-8a97-3b65f3254672\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.271958 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume\") pod \"4a150895-3ed5-4f19-8a97-3b65f3254672\" (UID: \"4a150895-3ed5-4f19-8a97-3b65f3254672\") " Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.272579 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume" (OuterVolumeSpecName: "config-volume") pod "4a150895-3ed5-4f19-8a97-3b65f3254672" (UID: "4a150895-3ed5-4f19-8a97-3b65f3254672"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.273160 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a150895-3ed5-4f19-8a97-3b65f3254672-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.279365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4a150895-3ed5-4f19-8a97-3b65f3254672" (UID: "4a150895-3ed5-4f19-8a97-3b65f3254672"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.279409 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865" (OuterVolumeSpecName: "kube-api-access-zz865") pod "4a150895-3ed5-4f19-8a97-3b65f3254672" (UID: "4a150895-3ed5-4f19-8a97-3b65f3254672"). InnerVolumeSpecName "kube-api-access-zz865". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.376228 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz865\" (UniqueName: \"kubernetes.io/projected/4a150895-3ed5-4f19-8a97-3b65f3254672-kube-api-access-zz865\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.376292 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a150895-3ed5-4f19-8a97-3b65f3254672-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.903044 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" event={"ID":"4a150895-3ed5-4f19-8a97-3b65f3254672","Type":"ContainerDied","Data":"1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440"} Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.903380 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1480a7f03b93c1de37bb75e6fb6e0a26c994052f0e7ff7ab7ed209cd18d75440" Mar 18 10:30:03 crc kubenswrapper[4733]: I0318 10:30:03.903138 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563830-kdgdh" Mar 18 10:30:06 crc kubenswrapper[4733]: I0318 10:30:06.825177 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-chmbd" Mar 18 10:30:07 crc kubenswrapper[4733]: I0318 10:30:07.116102 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-9txbj" Mar 18 10:30:07 crc kubenswrapper[4733]: I0318 10:30:07.174719 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-fd4t7" Mar 18 10:30:07 crc kubenswrapper[4733]: I0318 10:30:07.190094 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-v2pb2" Mar 18 10:30:07 crc kubenswrapper[4733]: I0318 10:30:07.254575 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" Mar 18 10:30:07 crc kubenswrapper[4733]: I0318 10:30:07.269810 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.359255 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.371352 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/651c7dd5-3adc-48b4-b579-309258aa3735-cert\") pod \"infra-operator-controller-manager-74c694b97b-j4snz\" (UID: \"651c7dd5-3adc-48b4-b579-309258aa3735\") " pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.529415 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vp5bk" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.537228 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.562976 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.570308 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6eca2f16-53b8-4173-ace4-18b7292b1369-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-s6rbv\" (UID: \"6eca2f16-53b8-4173-ace4-18b7292b1369\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.851420 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-gwtrb" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.858772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.978086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.978492 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.982916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-webhook-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.983017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4b7e706-a9a7-490a-84a8-094d1d909ba8-metrics-certs\") pod \"openstack-operator-controller-manager-85877db48-qvlf2\" (UID: \"a4b7e706-a9a7-490a-84a8-094d1d909ba8\") " pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:08 crc kubenswrapper[4733]: I0318 10:30:08.992932 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz"] Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.101264 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5xhfs" Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.109570 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.124339 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv"] Mar 18 10:30:09 crc kubenswrapper[4733]: W0318 10:30:09.131898 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eca2f16_53b8_4173_ace4_18b7292b1369.slice/crio-287488f351efea943c590067dadef0b34f1a7fcc14089ba9f594eddec06af49f WatchSource:0}: Error finding container 287488f351efea943c590067dadef0b34f1a7fcc14089ba9f594eddec06af49f: Status 404 returned error can't find the container with id 287488f351efea943c590067dadef0b34f1a7fcc14089ba9f594eddec06af49f Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.331415 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2"] Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.951388 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" containerID="be977da7d932bb787db9cafb1727d3bd50b5e03495d1e8a82c232ed7c66e241e" exitCode=0 Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.951449 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" event={"ID":"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35","Type":"ContainerDied","Data":"be977da7d932bb787db9cafb1727d3bd50b5e03495d1e8a82c232ed7c66e241e"} Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.953630 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" event={"ID":"6eca2f16-53b8-4173-ace4-18b7292b1369","Type":"ContainerStarted","Data":"287488f351efea943c590067dadef0b34f1a7fcc14089ba9f594eddec06af49f"} Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.956121 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" event={"ID":"a4b7e706-a9a7-490a-84a8-094d1d909ba8","Type":"ContainerStarted","Data":"fc4421f4ae5159781decebb7cafcfbdfcf0728a6d7062d996cbac8ffa23ff30d"} Mar 18 10:30:09 crc kubenswrapper[4733]: I0318 10:30:09.957560 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" event={"ID":"651c7dd5-3adc-48b4-b579-309258aa3735","Type":"ContainerStarted","Data":"12cfa674f121e9b5a9b930563f89df0d163c4e8f5b65a11351773057b72f3df1"} Mar 18 10:30:10 crc kubenswrapper[4733]: I0318 10:30:10.971523 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" event={"ID":"a4b7e706-a9a7-490a-84a8-094d1d909ba8","Type":"ContainerStarted","Data":"19a4ab4122ea3ab13731e02ddad2ee9bbfc4c18d66f55e966aed85c4621a2208"} Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.017133 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" podStartSLOduration=35.017106993 podStartE2EDuration="35.017106993s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:30:11.007416188 +0000 UTC m=+1050.499150533" watchObservedRunningTime="2026-03-18 10:30:11.017106993 +0000 UTC m=+1050.508841328" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.300976 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.416278 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz2gq\" (UniqueName: \"kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq\") pod \"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35\" (UID: \"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35\") " Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.424758 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq" (OuterVolumeSpecName: "kube-api-access-vz2gq") pod "3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" (UID: "3c8eb139-c576-4ab3-8c2b-a309f3aa4a35"). InnerVolumeSpecName "kube-api-access-vz2gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.518054 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz2gq\" (UniqueName: \"kubernetes.io/projected/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35-kube-api-access-vz2gq\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.986524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" event={"ID":"3c8eb139-c576-4ab3-8c2b-a309f3aa4a35","Type":"ContainerDied","Data":"8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765"} Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.989023 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d4e6c4fcb2ee48b7889361535947998ed4f0556343598ce1b623b89f6b9e765" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.989074 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:11 crc kubenswrapper[4733]: I0318 10:30:11.986598 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563830-2qqd2" Mar 18 10:30:12 crc kubenswrapper[4733]: I0318 10:30:12.386763 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563824-l49xk"] Mar 18 10:30:12 crc kubenswrapper[4733]: I0318 10:30:12.393088 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563824-l49xk"] Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.195665 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6077c15f-e285-4625-b336-a84327b1af2d" path="/var/lib/kubelet/pods/6077c15f-e285-4625-b336-a84327b1af2d/volumes" Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.570887 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.570956 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.571012 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.571635 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:30:13 crc kubenswrapper[4733]: I0318 10:30:13.571698 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96" gracePeriod=600 Mar 18 10:30:14 crc kubenswrapper[4733]: I0318 10:30:14.004404 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96" exitCode=0 Mar 18 10:30:14 crc kubenswrapper[4733]: I0318 10:30:14.004449 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96"} Mar 18 10:30:14 crc kubenswrapper[4733]: I0318 10:30:14.004485 4733 scope.go:117] "RemoveContainer" containerID="a11e956cdd33846b5919c35822db029436f82987d5e2c2bb6427c6d1dfd2098c" Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.012713 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625"} Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.015118 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" event={"ID":"6eca2f16-53b8-4173-ace4-18b7292b1369","Type":"ContainerStarted","Data":"61a53e6a37ef75cdfb62d9ca277be1f5b865c3fe9a1b65b59be71e8453a3d337"} Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.015536 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.016829 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" event={"ID":"651c7dd5-3adc-48b4-b579-309258aa3735","Type":"ContainerStarted","Data":"226fcfa3ddb6adc8e926f181f47c3dd1d4bd8a04122f509d852cfa38b5c8743f"} Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.017530 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.040503 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" podStartSLOduration=34.04694815 podStartE2EDuration="39.040482711s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:30:09.003245733 +0000 UTC m=+1048.494980058" lastFinishedPulling="2026-03-18 10:30:13.996780294 +0000 UTC m=+1053.488514619" observedRunningTime="2026-03-18 10:30:15.036694304 +0000 UTC m=+1054.528428629" watchObservedRunningTime="2026-03-18 10:30:15.040482711 +0000 UTC m=+1054.532217036" Mar 18 10:30:15 crc kubenswrapper[4733]: I0318 10:30:15.061023 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" podStartSLOduration=34.163728425 podStartE2EDuration="39.061005502s" podCreationTimestamp="2026-03-18 10:29:36 +0000 UTC" firstStartedPulling="2026-03-18 10:30:09.136810653 +0000 UTC m=+1048.628544988" lastFinishedPulling="2026-03-18 10:30:14.03408774 +0000 UTC m=+1053.525822065" observedRunningTime="2026-03-18 10:30:15.05917518 +0000 UTC m=+1054.550909505" watchObservedRunningTime="2026-03-18 10:30:15.061005502 +0000 UTC m=+1054.552739837" Mar 18 10:30:19 crc kubenswrapper[4733]: I0318 10:30:19.120856 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-85877db48-qvlf2" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.144001 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:28 crc kubenswrapper[4733]: E0318 10:30:28.144734 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a150895-3ed5-4f19-8a97-3b65f3254672" containerName="collect-profiles" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.144745 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a150895-3ed5-4f19-8a97-3b65f3254672" containerName="collect-profiles" Mar 18 10:30:28 crc kubenswrapper[4733]: E0318 10:30:28.144767 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" containerName="oc" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.144773 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" containerName="oc" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.144898 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a150895-3ed5-4f19-8a97-3b65f3254672" containerName="collect-profiles" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.144918 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" containerName="oc" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.145834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.162851 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.180704 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.180739 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42hs8\" (UniqueName: \"kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.180886 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.282342 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.282417 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.282437 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42hs8\" (UniqueName: \"kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.282768 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.283037 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.308577 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42hs8\" (UniqueName: \"kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8\") pod \"certified-operators-gxrnl\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.464484 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.544648 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-74c694b97b-j4snz" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.868075 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-s6rbv" Mar 18 10:30:28 crc kubenswrapper[4733]: I0318 10:30:28.967340 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:29 crc kubenswrapper[4733]: I0318 10:30:29.138288 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerStarted","Data":"3f798c2bf3f5a1669f683d8681c89d7676157ae54ed2fa5529e930b00288aa89"} Mar 18 10:30:30 crc kubenswrapper[4733]: I0318 10:30:30.151474 4733 generic.go:334] "Generic (PLEG): container finished" podID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerID="1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7" exitCode=0 Mar 18 10:30:30 crc kubenswrapper[4733]: I0318 10:30:30.151814 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerDied","Data":"1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7"} Mar 18 10:30:31 crc kubenswrapper[4733]: I0318 10:30:31.159371 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerStarted","Data":"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7"} Mar 18 10:30:32 crc kubenswrapper[4733]: I0318 10:30:32.169483 4733 generic.go:334] "Generic (PLEG): container finished" podID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerID="1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7" exitCode=0 Mar 18 10:30:32 crc kubenswrapper[4733]: I0318 10:30:32.169595 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerDied","Data":"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7"} Mar 18 10:30:33 crc kubenswrapper[4733]: I0318 10:30:33.202654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerStarted","Data":"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883"} Mar 18 10:30:33 crc kubenswrapper[4733]: I0318 10:30:33.214929 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gxrnl" podStartSLOduration=2.795983299 podStartE2EDuration="5.214906241s" podCreationTimestamp="2026-03-18 10:30:28 +0000 UTC" firstStartedPulling="2026-03-18 10:30:30.154466014 +0000 UTC m=+1069.646200379" lastFinishedPulling="2026-03-18 10:30:32.573388976 +0000 UTC m=+1072.065123321" observedRunningTime="2026-03-18 10:30:33.21241585 +0000 UTC m=+1072.704150205" watchObservedRunningTime="2026-03-18 10:30:33.214906241 +0000 UTC m=+1072.706640586" Mar 18 10:30:38 crc kubenswrapper[4733]: I0318 10:30:38.465685 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:38 crc kubenswrapper[4733]: I0318 10:30:38.465762 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:38 crc kubenswrapper[4733]: I0318 10:30:38.523711 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:39 crc kubenswrapper[4733]: I0318 10:30:39.300405 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:39 crc kubenswrapper[4733]: I0318 10:30:39.346878 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.252513 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gxrnl" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="registry-server" containerID="cri-o://88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883" gracePeriod=2 Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.711021 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.767771 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities\") pod \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.767955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42hs8\" (UniqueName: \"kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8\") pod \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.768029 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content\") pod \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\" (UID: \"be59cd8b-2e8d-41ae-8c18-1f6dea878859\") " Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.768901 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities" (OuterVolumeSpecName: "utilities") pod "be59cd8b-2e8d-41ae-8c18-1f6dea878859" (UID: "be59cd8b-2e8d-41ae-8c18-1f6dea878859"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.775030 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8" (OuterVolumeSpecName: "kube-api-access-42hs8") pod "be59cd8b-2e8d-41ae-8c18-1f6dea878859" (UID: "be59cd8b-2e8d-41ae-8c18-1f6dea878859"). InnerVolumeSpecName "kube-api-access-42hs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.831296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be59cd8b-2e8d-41ae-8c18-1f6dea878859" (UID: "be59cd8b-2e8d-41ae-8c18-1f6dea878859"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.869722 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42hs8\" (UniqueName: \"kubernetes.io/projected/be59cd8b-2e8d-41ae-8c18-1f6dea878859-kube-api-access-42hs8\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.869768 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:41 crc kubenswrapper[4733]: I0318 10:30:41.869781 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be59cd8b-2e8d-41ae-8c18-1f6dea878859-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.260089 4733 generic.go:334] "Generic (PLEG): container finished" podID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerID="88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883" exitCode=0 Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.260138 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gxrnl" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.260157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerDied","Data":"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883"} Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.260641 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gxrnl" event={"ID":"be59cd8b-2e8d-41ae-8c18-1f6dea878859","Type":"ContainerDied","Data":"3f798c2bf3f5a1669f683d8681c89d7676157ae54ed2fa5529e930b00288aa89"} Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.260673 4733 scope.go:117] "RemoveContainer" containerID="88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.288580 4733 scope.go:117] "RemoveContainer" containerID="1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.297513 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.303826 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gxrnl"] Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.309626 4733 scope.go:117] "RemoveContainer" containerID="1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.330287 4733 scope.go:117] "RemoveContainer" containerID="88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883" Mar 18 10:30:42 crc kubenswrapper[4733]: E0318 10:30:42.330702 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883\": container with ID starting with 88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883 not found: ID does not exist" containerID="88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.330739 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883"} err="failed to get container status \"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883\": rpc error: code = NotFound desc = could not find container \"88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883\": container with ID starting with 88d1674a4eaf402acf29785a0f8030272e4ba1e43ad27f542258f3ccecd14883 not found: ID does not exist" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.330764 4733 scope.go:117] "RemoveContainer" containerID="1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7" Mar 18 10:30:42 crc kubenswrapper[4733]: E0318 10:30:42.331030 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7\": container with ID starting with 1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7 not found: ID does not exist" containerID="1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.331078 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7"} err="failed to get container status \"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7\": rpc error: code = NotFound desc = could not find container \"1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7\": container with ID starting with 1f2ad3b9ebbdb27ca1e9ac77090b1c05e9f6528d3dbbd9391a8e464389617db7 not found: ID does not exist" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.331109 4733 scope.go:117] "RemoveContainer" containerID="1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7" Mar 18 10:30:42 crc kubenswrapper[4733]: E0318 10:30:42.331454 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7\": container with ID starting with 1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7 not found: ID does not exist" containerID="1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7" Mar 18 10:30:42 crc kubenswrapper[4733]: I0318 10:30:42.331489 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7"} err="failed to get container status \"1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7\": rpc error: code = NotFound desc = could not find container \"1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7\": container with ID starting with 1a0d3f9f18feb5d8d80a2b6645f85d5b1828acf7d2348ee16404888b26c9e7d7 not found: ID does not exist" Mar 18 10:30:43 crc kubenswrapper[4733]: I0318 10:30:43.194475 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" path="/var/lib/kubelet/pods/be59cd8b-2e8d-41ae-8c18-1f6dea878859/volumes" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.206646 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:30:48 crc kubenswrapper[4733]: E0318 10:30:48.208587 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="registry-server" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.208719 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="registry-server" Mar 18 10:30:48 crc kubenswrapper[4733]: E0318 10:30:48.208805 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="extract-utilities" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.208875 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="extract-utilities" Mar 18 10:30:48 crc kubenswrapper[4733]: E0318 10:30:48.208965 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="extract-content" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.209040 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="extract-content" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.209312 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="be59cd8b-2e8d-41ae-8c18-1f6dea878859" containerName="registry-server" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.210462 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.215723 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.219151 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.219554 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.223940 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-c44sb" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.224227 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.296625 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.297821 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.309370 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.324057 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.379895 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.380027 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtr2d\" (UniqueName: \"kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.480973 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.481029 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtr2d\" (UniqueName: \"kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.481114 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.481141 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.481211 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc6zn\" (UniqueName: \"kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.482170 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.503916 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtr2d\" (UniqueName: \"kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d\") pod \"dnsmasq-dns-675f4bcbfc-7hxp6\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.535656 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.582024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.582702 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc6zn\" (UniqueName: \"kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.582871 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.583032 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.583479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.619106 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc6zn\" (UniqueName: \"kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn\") pod \"dnsmasq-dns-78dd6ddcc-jhs6c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.623502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:30:48 crc kubenswrapper[4733]: I0318 10:30:48.932811 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:30:49 crc kubenswrapper[4733]: I0318 10:30:49.051242 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:30:49 crc kubenswrapper[4733]: W0318 10:30:49.052580 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod681a4fb9_f5dc_4b7d_aad7_45d15f11de1c.slice/crio-143eafdd483ed64fb00fabfeb37bfb5824c7d4e108597ececea0f80ea29d068b WatchSource:0}: Error finding container 143eafdd483ed64fb00fabfeb37bfb5824c7d4e108597ececea0f80ea29d068b: Status 404 returned error can't find the container with id 143eafdd483ed64fb00fabfeb37bfb5824c7d4e108597ececea0f80ea29d068b Mar 18 10:30:49 crc kubenswrapper[4733]: I0318 10:30:49.319540 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" event={"ID":"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c","Type":"ContainerStarted","Data":"143eafdd483ed64fb00fabfeb37bfb5824c7d4e108597ececea0f80ea29d068b"} Mar 18 10:30:49 crc kubenswrapper[4733]: I0318 10:30:49.321828 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" event={"ID":"725b76ca-c6aa-47f4-b75b-7ba4cd999979","Type":"ContainerStarted","Data":"338afcfd61f57729b6823a0742441ae38ca9d73a3a0cb99ea518e0556be8e8f6"} Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.039456 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.059932 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.061177 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.095895 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.231689 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.231749 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mfbr\" (UniqueName: \"kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.231840 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.333765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.333838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.333853 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mfbr\" (UniqueName: \"kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.334852 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.335342 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.374346 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.389149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mfbr\" (UniqueName: \"kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr\") pod \"dnsmasq-dns-666b6646f7-xh24t\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.389524 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.398691 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.400006 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.410493 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.535855 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tggzq\" (UniqueName: \"kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.536237 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.536276 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.637737 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tggzq\" (UniqueName: \"kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.637864 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.637930 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.638800 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.638858 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.661261 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tggzq\" (UniqueName: \"kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq\") pod \"dnsmasq-dns-57d769cc4f-tljb4\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.741591 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:30:51 crc kubenswrapper[4733]: I0318 10:30:51.975908 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.012459 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.013637 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017367 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017386 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017685 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017760 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p7fvd" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017857 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.017990 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.018063 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.023618 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wcnb\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-kube-api-access-7wcnb\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145374 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145411 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145439 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145459 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0570ce4-1455-4698-85cf-01f7108d9e7f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145510 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145524 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0570ce4-1455-4698-85cf-01f7108d9e7f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.145556 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.146368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.179228 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247397 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247438 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0570ce4-1455-4698-85cf-01f7108d9e7f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247457 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247501 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wcnb\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-kube-api-access-7wcnb\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247560 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247611 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247627 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.247653 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0570ce4-1455-4698-85cf-01f7108d9e7f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.248644 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.248935 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.248949 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.250814 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.251140 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.251333 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0570ce4-1455-4698-85cf-01f7108d9e7f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.260037 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.260113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0570ce4-1455-4698-85cf-01f7108d9e7f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.260134 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0570ce4-1455-4698-85cf-01f7108d9e7f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.260236 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.265393 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wcnb\" (UniqueName: \"kubernetes.io/projected/f0570ce4-1455-4698-85cf-01f7108d9e7f-kube-api-access-7wcnb\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.268950 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"f0570ce4-1455-4698-85cf-01f7108d9e7f\") " pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.341430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.360015 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.363572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366317 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366644 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366729 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366788 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366901 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366920 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6884w" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.366921 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.376399 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.551941 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552002 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s62ng\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-kube-api-access-s62ng\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552082 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552121 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552168 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552225 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552273 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552315 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.552350 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654752 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s62ng\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-kube-api-access-s62ng\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654827 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654886 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654965 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.654999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.655045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.655083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.655123 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.656513 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.657377 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.657830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.658509 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.659817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.660307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.661716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.662450 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.657535 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.674113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.678485 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s62ng\" (UniqueName: \"kubernetes.io/projected/b4a4e3e2-bd4d-4f8d-97bc-51267378ab03-kube-api-access-s62ng\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.684955 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:52 crc kubenswrapper[4733]: I0318 10:30:52.713812 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.565845 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.566919 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.575028 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.577508 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.577652 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-7hltt" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.577754 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.577837 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.584584 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691394 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-operator-scripts\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691462 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kolla-config\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691615 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-default\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691699 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-generated\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691742 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691768 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjj5v\" (UniqueName: \"kubernetes.io/projected/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kube-api-access-zjj5v\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691828 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.691893 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.792966 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793036 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-operator-scripts\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kolla-config\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793128 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-default\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793169 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-generated\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793212 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793237 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjj5v\" (UniqueName: \"kubernetes.io/projected/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kube-api-access-zjj5v\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.793276 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.794718 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-default\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.796031 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-operator-scripts\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.796613 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kolla-config\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.796895 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.802886 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.803704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-config-data-generated\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.821507 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.823815 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjj5v\" (UniqueName: \"kubernetes.io/projected/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-kube-api-access-zjj5v\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.828798 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc60b49b-96fa-40fd-a8e5-40c810f5ef80-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"dc60b49b-96fa-40fd-a8e5-40c810f5ef80\") " pod="openstack/openstack-galera-0" Mar 18 10:30:53 crc kubenswrapper[4733]: I0318 10:30:53.906550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 18 10:30:54 crc kubenswrapper[4733]: I0318 10:30:54.942179 4733 scope.go:117] "RemoveContainer" containerID="6f50555f9faf96f94c8c33f53803364eb9620cbe1dd5e27e68cba9056a299fa1" Mar 18 10:30:54 crc kubenswrapper[4733]: I0318 10:30:54.999716 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.000826 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.002443 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-67rtq" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.002819 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.002865 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.003074 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.017403 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130385 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130450 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjknb\" (UniqueName: \"kubernetes.io/projected/0208d826-df0f-41c8-83a7-821a21b7b85d-kube-api-access-cjknb\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130481 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130506 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130551 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130606 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130643 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.130670 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.132482 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.133536 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.135753 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.135920 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dssff" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.136033 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.144455 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232281 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232341 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-config-data\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232377 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232403 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232448 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232502 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvwbt\" (UniqueName: \"kubernetes.io/projected/dd66892e-808c-405a-ac8e-366b6ca8b148-kube-api-access-kvwbt\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232523 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232554 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232577 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-kolla-config\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232723 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.232785 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjknb\" (UniqueName: \"kubernetes.io/projected/0208d826-df0f-41c8-83a7-821a21b7b85d-kube-api-access-cjknb\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.233009 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.233051 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.233919 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.234100 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.234160 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0208d826-df0f-41c8-83a7-821a21b7b85d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.234704 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0208d826-df0f-41c8-83a7-821a21b7b85d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.238129 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.238729 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0208d826-df0f-41c8-83a7-821a21b7b85d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.256525 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjknb\" (UniqueName: \"kubernetes.io/projected/0208d826-df0f-41c8-83a7-821a21b7b85d-kube-api-access-cjknb\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.263868 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0208d826-df0f-41c8-83a7-821a21b7b85d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.320429 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.335170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvwbt\" (UniqueName: \"kubernetes.io/projected/dd66892e-808c-405a-ac8e-366b6ca8b148-kube-api-access-kvwbt\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.335235 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.335256 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-kolla-config\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.335325 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-config-data\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.335706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.336463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-kolla-config\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.336498 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd66892e-808c-405a-ac8e-366b6ca8b148-config-data\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.338222 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.338520 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd66892e-808c-405a-ac8e-366b6ca8b148-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.352246 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvwbt\" (UniqueName: \"kubernetes.io/projected/dd66892e-808c-405a-ac8e-366b6ca8b148-kube-api-access-kvwbt\") pod \"memcached-0\" (UID: \"dd66892e-808c-405a-ac8e-366b6ca8b148\") " pod="openstack/memcached-0" Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.392583 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" event={"ID":"a56bac49-b398-4b61-9b54-7969acd2dc93","Type":"ContainerStarted","Data":"cf4476f986138503b3408c91ec78e55f73e59536bc0804b03b95667a22e6c6a6"} Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.394341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" event={"ID":"d6ec9568-99c8-4bee-a97c-46400fcc0e73","Type":"ContainerStarted","Data":"642fd85cd38aa86a0c841253dfb5fda87ce7251b22474e12b3ef33923f92f9b2"} Mar 18 10:30:55 crc kubenswrapper[4733]: I0318 10:30:55.456704 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.368389 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.373403 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.376089 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.377484 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-mzkrd" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.401853 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77qsg\" (UniqueName: \"kubernetes.io/projected/55f0b230-09f2-4be2-aa1f-76a37f3fe30c-kube-api-access-77qsg\") pod \"kube-state-metrics-0\" (UID: \"55f0b230-09f2-4be2-aa1f-76a37f3fe30c\") " pod="openstack/kube-state-metrics-0" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.503317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77qsg\" (UniqueName: \"kubernetes.io/projected/55f0b230-09f2-4be2-aa1f-76a37f3fe30c-kube-api-access-77qsg\") pod \"kube-state-metrics-0\" (UID: \"55f0b230-09f2-4be2-aa1f-76a37f3fe30c\") " pod="openstack/kube-state-metrics-0" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.523214 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77qsg\" (UniqueName: \"kubernetes.io/projected/55f0b230-09f2-4be2-aa1f-76a37f3fe30c-kube-api-access-77qsg\") pod \"kube-state-metrics-0\" (UID: \"55f0b230-09f2-4be2-aa1f-76a37f3fe30c\") " pod="openstack/kube-state-metrics-0" Mar 18 10:30:57 crc kubenswrapper[4733]: I0318 10:30:57.707762 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.482283 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rh64b"] Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.483502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.487472 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.487519 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.487769 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-84shn" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.492440 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-ljrgt"] Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.493915 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.500230 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rh64b"] Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.512798 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ljrgt"] Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553253 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-run\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q7j4\" (UniqueName: \"kubernetes.io/projected/d75a8d54-aca8-49cd-9062-6389baaf7a09-kube-api-access-9q7j4\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553424 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-scripts\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553478 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-etc-ovs\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-lib\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553553 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-combined-ca-bundle\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdpd5\" (UniqueName: \"kubernetes.io/projected/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-kube-api-access-sdpd5\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553612 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-log-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553653 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-log\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553741 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553763 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75a8d54-aca8-49cd-9062-6389baaf7a09-scripts\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553788 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-ovn-controller-tls-certs\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.553858 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654607 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-run\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q7j4\" (UniqueName: \"kubernetes.io/projected/d75a8d54-aca8-49cd-9062-6389baaf7a09-kube-api-access-9q7j4\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-scripts\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654705 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-etc-ovs\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654730 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-lib\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654748 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdpd5\" (UniqueName: \"kubernetes.io/projected/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-kube-api-access-sdpd5\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-combined-ca-bundle\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654783 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-log-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654806 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-log\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654830 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75a8d54-aca8-49cd-9062-6389baaf7a09-scripts\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654863 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-ovn-controller-tls-certs\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.654894 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.655221 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.655239 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-run\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.656166 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-log-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.656265 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-log\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.656343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-var-run-ovn\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.657885 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-etc-ovs\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.658923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d75a8d54-aca8-49cd-9062-6389baaf7a09-var-lib\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.660422 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75a8d54-aca8-49cd-9062-6389baaf7a09-scripts\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.663986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-scripts\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.673231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-ovn-controller-tls-certs\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.673290 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q7j4\" (UniqueName: \"kubernetes.io/projected/d75a8d54-aca8-49cd-9062-6389baaf7a09-kube-api-access-9q7j4\") pod \"ovn-controller-ovs-ljrgt\" (UID: \"d75a8d54-aca8-49cd-9062-6389baaf7a09\") " pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.673720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-combined-ca-bundle\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.677216 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdpd5\" (UniqueName: \"kubernetes.io/projected/e3c842d3-b3dd-4cf2-9df0-16cea4061bc5-kube-api-access-sdpd5\") pod \"ovn-controller-rh64b\" (UID: \"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5\") " pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.837847 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b" Mar 18 10:31:00 crc kubenswrapper[4733]: I0318 10:31:00.849038 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.391927 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.393284 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.396222 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.396730 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.396777 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rnxnj" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.396870 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.396933 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.399270 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.569919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95hf9\" (UniqueName: \"kubernetes.io/projected/a8c27598-870d-4de0-a986-47042d7d6f4c-kube-api-access-95hf9\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570005 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570080 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570166 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570220 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570246 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-config\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.570457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672578 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672648 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672706 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95hf9\" (UniqueName: \"kubernetes.io/projected/a8c27598-870d-4de0-a986-47042d7d6f4c-kube-api-access-95hf9\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672794 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.672881 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-config\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.673347 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.678650 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.680452 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.680953 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.681631 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8c27598-870d-4de0-a986-47042d7d6f4c-config\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.688878 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.690574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95hf9\" (UniqueName: \"kubernetes.io/projected/a8c27598-870d-4de0-a986-47042d7d6f4c-kube-api-access-95hf9\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.695122 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c27598-870d-4de0-a986-47042d7d6f4c-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.698126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a8c27598-870d-4de0-a986-47042d7d6f4c\") " pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:01 crc kubenswrapper[4733]: I0318 10:31:01.720732 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.867979 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.868483 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mtr2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-7hxp6_openstack(725b76ca-c6aa-47f4-b75b-7ba4cd999979): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.869696 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" podUID="725b76ca-c6aa-47f4-b75b-7ba4cd999979" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.915455 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.915611 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dc6zn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-jhs6c_openstack(681a4fb9-f5dc-4b7d-aad7-45d15f11de1c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 10:31:03 crc kubenswrapper[4733]: E0318 10:31:03.916770 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" podUID="681a4fb9-f5dc-4b7d-aad7-45d15f11de1c" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.480869 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerID="f1bfaec48682e00a29092241786f91d62461cdabac1600bdef26a808c7697bdd" exitCode=0 Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.482241 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" event={"ID":"d6ec9568-99c8-4bee-a97c-46400fcc0e73","Type":"ContainerDied","Data":"f1bfaec48682e00a29092241786f91d62461cdabac1600bdef26a808c7697bdd"} Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.485740 4733 generic.go:334] "Generic (PLEG): container finished" podID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerID="0d7f9ebe26354b0fdcafbf9243319e234596728fe76f62846f8d6f2de9c01686" exitCode=0 Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.487035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" event={"ID":"a56bac49-b398-4b61-9b54-7969acd2dc93","Type":"ContainerDied","Data":"0d7f9ebe26354b0fdcafbf9243319e234596728fe76f62846f8d6f2de9c01686"} Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.570056 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.571633 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.576294 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.576509 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.576646 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-726jj" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.578596 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.593503 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.597649 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.605567 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.623674 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 10:31:04 crc kubenswrapper[4733]: W0318 10:31:04.666906 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc60b49b_96fa_40fd_a8e5_40c810f5ef80.slice/crio-1cc1d503cfe0c3f0293efb60f8e37ebd8029bc5f9e2559b027fd01da9fcfc135 WatchSource:0}: Error finding container 1cc1d503cfe0c3f0293efb60f8e37ebd8029bc5f9e2559b027fd01da9fcfc135: Status 404 returned error can't find the container with id 1cc1d503cfe0c3f0293efb60f8e37ebd8029bc5f9e2559b027fd01da9fcfc135 Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.720820 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ljrgt"] Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760656 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-config\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760705 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760826 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760858 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760904 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-842ht\" (UniqueName: \"kubernetes.io/projected/0868210e-9d93-4f63-b425-7db21f13cd90-kube-api-access-842ht\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.760971 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.862636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-842ht\" (UniqueName: \"kubernetes.io/projected/0868210e-9d93-4f63-b425-7db21f13cd90-kube-api-access-842ht\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863090 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863130 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-config\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863164 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863262 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863316 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863346 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863399 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.863901 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.864017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-config\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.864048 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.866791 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0868210e-9d93-4f63-b425-7db21f13cd90-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.869219 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.869253 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.873958 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0868210e-9d93-4f63-b425-7db21f13cd90-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.882042 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.883322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-842ht\" (UniqueName: \"kubernetes.io/projected/0868210e-9d93-4f63-b425-7db21f13cd90-kube-api-access-842ht\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.893463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0868210e-9d93-4f63-b425-7db21f13cd90\") " pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.893842 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.903698 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:04 crc kubenswrapper[4733]: I0318 10:31:04.977593 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 10:31:05 crc kubenswrapper[4733]: W0318 10:31:05.005947 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0208d826_df0f_41c8_83a7_821a21b7b85d.slice/crio-c832fa08ef6fe02ea7fd8206a78bd9fcd5b47ee8734823fb799cd2309fa0a789 WatchSource:0}: Error finding container c832fa08ef6fe02ea7fd8206a78bd9fcd5b47ee8734823fb799cd2309fa0a789: Status 404 returned error can't find the container with id c832fa08ef6fe02ea7fd8206a78bd9fcd5b47ee8734823fb799cd2309fa0a789 Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.019153 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rh64b"] Mar 18 10:31:05 crc kubenswrapper[4733]: W0318 10:31:05.030448 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3c842d3_b3dd_4cf2_9df0_16cea4061bc5.slice/crio-b6ddef781f663dec8631c0b8e7931ac50c253872640374c5f5e682ce10a5bdce WatchSource:0}: Error finding container b6ddef781f663dec8631c0b8e7931ac50c253872640374c5f5e682ce10a5bdce: Status 404 returned error can't find the container with id b6ddef781f663dec8631c0b8e7931ac50c253872640374c5f5e682ce10a5bdce Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.043587 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 10:31:05 crc kubenswrapper[4733]: W0318 10:31:05.060749 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4a4e3e2_bd4d_4f8d_97bc_51267378ab03.slice/crio-b52c27ec7e9f93d315a6dda8b5e47a3247a65d8221d7f759cf1532f592f883b1 WatchSource:0}: Error finding container b52c27ec7e9f93d315a6dda8b5e47a3247a65d8221d7f759cf1532f592f883b1: Status 404 returned error can't find the container with id b52c27ec7e9f93d315a6dda8b5e47a3247a65d8221d7f759cf1532f592f883b1 Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.064559 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.067860 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc\") pod \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.067995 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc6zn\" (UniqueName: \"kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn\") pod \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.068034 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config\") pod \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.068118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config\") pod \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\" (UID: \"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c\") " Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.068154 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtr2d\" (UniqueName: \"kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d\") pod \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\" (UID: \"725b76ca-c6aa-47f4-b75b-7ba4cd999979\") " Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.069211 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c" (UID: "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.069597 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config" (OuterVolumeSpecName: "config") pod "725b76ca-c6aa-47f4-b75b-7ba4cd999979" (UID: "725b76ca-c6aa-47f4-b75b-7ba4cd999979"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.069650 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config" (OuterVolumeSpecName: "config") pod "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c" (UID: "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.074154 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d" (OuterVolumeSpecName: "kube-api-access-mtr2d") pod "725b76ca-c6aa-47f4-b75b-7ba4cd999979" (UID: "725b76ca-c6aa-47f4-b75b-7ba4cd999979"). InnerVolumeSpecName "kube-api-access-mtr2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.075044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn" (OuterVolumeSpecName: "kube-api-access-dc6zn") pod "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c" (UID: "681a4fb9-f5dc-4b7d-aad7-45d15f11de1c"). InnerVolumeSpecName "kube-api-access-dc6zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.171309 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.171693 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtr2d\" (UniqueName: \"kubernetes.io/projected/725b76ca-c6aa-47f4-b75b-7ba4cd999979-kube-api-access-mtr2d\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.171711 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.171724 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc6zn\" (UniqueName: \"kubernetes.io/projected/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c-kube-api-access-dc6zn\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.171763 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/725b76ca-c6aa-47f4-b75b-7ba4cd999979-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.429284 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.495518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ljrgt" event={"ID":"d75a8d54-aca8-49cd-9062-6389baaf7a09","Type":"ContainerStarted","Data":"f6d52325fe24512c4cbe1a8ba8f41b25ad326a0a8ebc7201acee3673f1955213"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.498212 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" event={"ID":"a56bac49-b398-4b61-9b54-7969acd2dc93","Type":"ContainerStarted","Data":"9ab9c30c68bdbce8904477ffce48bb66e4703ed2c3ced83026789c4904ce7735"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.499126 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.500440 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rh64b" event={"ID":"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5","Type":"ContainerStarted","Data":"b6ddef781f663dec8631c0b8e7931ac50c253872640374c5f5e682ce10a5bdce"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.502412 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"5e84fda0307415591d2c9d4daad4d37b5e33c659c7910fb7b6abb0d132644f7d"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.503335 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55f0b230-09f2-4be2-aa1f-76a37f3fe30c","Type":"ContainerStarted","Data":"668c3f35174c4e84be42265f1c88bf5ad344602872ac5d7aa3490f7a7785d7a2"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.504773 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" event={"ID":"681a4fb9-f5dc-4b7d-aad7-45d15f11de1c","Type":"ContainerDied","Data":"143eafdd483ed64fb00fabfeb37bfb5824c7d4e108597ececea0f80ea29d068b"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.504827 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jhs6c" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.506925 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"b52c27ec7e9f93d315a6dda8b5e47a3247a65d8221d7f759cf1532f592f883b1"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.508302 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0208d826-df0f-41c8-83a7-821a21b7b85d","Type":"ContainerStarted","Data":"c832fa08ef6fe02ea7fd8206a78bd9fcd5b47ee8734823fb799cd2309fa0a789"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.510341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"dc60b49b-96fa-40fd-a8e5-40c810f5ef80","Type":"ContainerStarted","Data":"1cc1d503cfe0c3f0293efb60f8e37ebd8029bc5f9e2559b027fd01da9fcfc135"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.517344 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" podStartSLOduration=5.193686977 podStartE2EDuration="14.517325795s" podCreationTimestamp="2026-03-18 10:30:51 +0000 UTC" firstStartedPulling="2026-03-18 10:30:54.697752813 +0000 UTC m=+1094.189487138" lastFinishedPulling="2026-03-18 10:31:04.021391631 +0000 UTC m=+1103.513125956" observedRunningTime="2026-03-18 10:31:05.513399284 +0000 UTC m=+1105.005133619" watchObservedRunningTime="2026-03-18 10:31:05.517325795 +0000 UTC m=+1105.009060120" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.522326 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" event={"ID":"725b76ca-c6aa-47f4-b75b-7ba4cd999979","Type":"ContainerDied","Data":"338afcfd61f57729b6823a0742441ae38ca9d73a3a0cb99ea518e0556be8e8f6"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.522390 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7hxp6" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.526954 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd66892e-808c-405a-ac8e-366b6ca8b148","Type":"ContainerStarted","Data":"815f9c448060d19c188316891f4b7a7913171ca8b6609a7cca214bade89395e1"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.532261 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" event={"ID":"d6ec9568-99c8-4bee-a97c-46400fcc0e73","Type":"ContainerStarted","Data":"aef65315ca9ffdfc8f038573ba4d7c7dad7e497bb3b3225251bd4117319d3146"} Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.537148 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.649501 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.668276 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jhs6c"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.684418 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" podStartSLOduration=5.321941026 podStartE2EDuration="14.684397223s" podCreationTimestamp="2026-03-18 10:30:51 +0000 UTC" firstStartedPulling="2026-03-18 10:30:54.699721529 +0000 UTC m=+1094.191455854" lastFinishedPulling="2026-03-18 10:31:04.062177716 +0000 UTC m=+1103.553912051" observedRunningTime="2026-03-18 10:31:05.551003458 +0000 UTC m=+1105.042737793" watchObservedRunningTime="2026-03-18 10:31:05.684397223 +0000 UTC m=+1105.176131548" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.721787 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.732931 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7hxp6"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.751383 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.935608 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-6trms"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.937279 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.939383 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 18 10:31:05 crc kubenswrapper[4733]: W0318 10:31:05.939577 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8c27598_870d_4de0_a986_47042d7d6f4c.slice/crio-b06a5b369913b1df1835f783d2fda94d69d9748e0d8b40519583ecfe90984a75 WatchSource:0}: Error finding container b06a5b369913b1df1835f783d2fda94d69d9748e0d8b40519583ecfe90984a75: Status 404 returned error can't find the container with id b06a5b369913b1df1835f783d2fda94d69d9748e0d8b40519583ecfe90984a75 Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.955900 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6trms"] Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-config\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998107 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998232 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovn-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998257 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-combined-ca-bundle\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998294 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovs-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:05 crc kubenswrapper[4733]: I0318 10:31:05.998397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h7tb\" (UniqueName: \"kubernetes.io/projected/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-kube-api-access-2h7tb\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.073848 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126785 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovn-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126828 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-combined-ca-bundle\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126847 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovs-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126916 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h7tb\" (UniqueName: \"kubernetes.io/projected/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-kube-api-access-2h7tb\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126940 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-config\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.126961 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.130209 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovs-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.130320 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-ovn-rundir\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.131131 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-config\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.137780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.141272 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-combined-ca-bundle\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.165518 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.166683 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.175455 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.194743 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h7tb\" (UniqueName: \"kubernetes.io/projected/e7849feb-5f1b-4b67-a3f7-8a419ebda0bd-kube-api-access-2h7tb\") pod \"ovn-controller-metrics-6trms\" (UID: \"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd\") " pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.203700 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.227961 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.228003 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.228048 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.228083 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrzpq\" (UniqueName: \"kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.307321 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.308820 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6trms" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.333231 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.333285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.333333 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.333367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrzpq\" (UniqueName: \"kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.334598 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.334771 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.337585 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.349091 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.349977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.351593 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.355918 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.369799 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrzpq\" (UniqueName: \"kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq\") pod \"dnsmasq-dns-5bf47b49b7-z55fc\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.436074 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.436153 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.436248 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5w85\" (UniqueName: \"kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.436274 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.436375 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.538280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.538352 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.538397 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5w85\" (UniqueName: \"kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.538422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.538496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.539945 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.540265 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.541580 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.543112 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.547475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0868210e-9d93-4f63-b425-7db21f13cd90","Type":"ContainerStarted","Data":"ee032d9694b4a39ad832b4921248ce94a5c6074e3c7ce3055076e87f6505375c"} Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.550686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a8c27598-870d-4de0-a986-47042d7d6f4c","Type":"ContainerStarted","Data":"b06a5b369913b1df1835f783d2fda94d69d9748e0d8b40519583ecfe90984a75"} Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.555118 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5w85\" (UniqueName: \"kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85\") pod \"dnsmasq-dns-8554648995-rrvg6\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.597212 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:06 crc kubenswrapper[4733]: I0318 10:31:06.668153 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:07 crc kubenswrapper[4733]: I0318 10:31:07.187504 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="681a4fb9-f5dc-4b7d-aad7-45d15f11de1c" path="/var/lib/kubelet/pods/681a4fb9-f5dc-4b7d-aad7-45d15f11de1c/volumes" Mar 18 10:31:07 crc kubenswrapper[4733]: I0318 10:31:07.188073 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="725b76ca-c6aa-47f4-b75b-7ba4cd999979" path="/var/lib/kubelet/pods/725b76ca-c6aa-47f4-b75b-7ba4cd999979/volumes" Mar 18 10:31:07 crc kubenswrapper[4733]: I0318 10:31:07.312680 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6trms"] Mar 18 10:31:07 crc kubenswrapper[4733]: I0318 10:31:07.561118 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="dnsmasq-dns" containerID="cri-o://9ab9c30c68bdbce8904477ffce48bb66e4703ed2c3ced83026789c4904ce7735" gracePeriod=10 Mar 18 10:31:07 crc kubenswrapper[4733]: I0318 10:31:07.561558 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="dnsmasq-dns" containerID="cri-o://aef65315ca9ffdfc8f038573ba4d7c7dad7e497bb3b3225251bd4117319d3146" gracePeriod=10 Mar 18 10:31:08 crc kubenswrapper[4733]: I0318 10:31:08.569393 4733 generic.go:334] "Generic (PLEG): container finished" podID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerID="9ab9c30c68bdbce8904477ffce48bb66e4703ed2c3ced83026789c4904ce7735" exitCode=0 Mar 18 10:31:08 crc kubenswrapper[4733]: I0318 10:31:08.569471 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" event={"ID":"a56bac49-b398-4b61-9b54-7969acd2dc93","Type":"ContainerDied","Data":"9ab9c30c68bdbce8904477ffce48bb66e4703ed2c3ced83026789c4904ce7735"} Mar 18 10:31:08 crc kubenswrapper[4733]: I0318 10:31:08.571976 4733 generic.go:334] "Generic (PLEG): container finished" podID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerID="aef65315ca9ffdfc8f038573ba4d7c7dad7e497bb3b3225251bd4117319d3146" exitCode=0 Mar 18 10:31:08 crc kubenswrapper[4733]: I0318 10:31:08.572053 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" event={"ID":"d6ec9568-99c8-4bee-a97c-46400fcc0e73","Type":"ContainerDied","Data":"aef65315ca9ffdfc8f038573ba4d7c7dad7e497bb3b3225251bd4117319d3146"} Mar 18 10:31:08 crc kubenswrapper[4733]: I0318 10:31:08.573021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6trms" event={"ID":"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd","Type":"ContainerStarted","Data":"e7e81fa094228e3b5ea6026dec0e1023cf195bd21cad69e71c8bbeda57a23101"} Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.101360 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.105454 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248303 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc\") pod \"a56bac49-b398-4b61-9b54-7969acd2dc93\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248348 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc\") pod \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248405 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config\") pod \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248431 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tggzq\" (UniqueName: \"kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq\") pod \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\" (UID: \"d6ec9568-99c8-4bee-a97c-46400fcc0e73\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248447 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config\") pod \"a56bac49-b398-4b61-9b54-7969acd2dc93\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.248469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mfbr\" (UniqueName: \"kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr\") pod \"a56bac49-b398-4b61-9b54-7969acd2dc93\" (UID: \"a56bac49-b398-4b61-9b54-7969acd2dc93\") " Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.261552 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr" (OuterVolumeSpecName: "kube-api-access-5mfbr") pod "a56bac49-b398-4b61-9b54-7969acd2dc93" (UID: "a56bac49-b398-4b61-9b54-7969acd2dc93"). InnerVolumeSpecName "kube-api-access-5mfbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.264272 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq" (OuterVolumeSpecName: "kube-api-access-tggzq") pod "d6ec9568-99c8-4bee-a97c-46400fcc0e73" (UID: "d6ec9568-99c8-4bee-a97c-46400fcc0e73"). InnerVolumeSpecName "kube-api-access-tggzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.304072 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config" (OuterVolumeSpecName: "config") pod "d6ec9568-99c8-4bee-a97c-46400fcc0e73" (UID: "d6ec9568-99c8-4bee-a97c-46400fcc0e73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.309069 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config" (OuterVolumeSpecName: "config") pod "a56bac49-b398-4b61-9b54-7969acd2dc93" (UID: "a56bac49-b398-4b61-9b54-7969acd2dc93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.322859 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6ec9568-99c8-4bee-a97c-46400fcc0e73" (UID: "d6ec9568-99c8-4bee-a97c-46400fcc0e73"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.323884 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a56bac49-b398-4b61-9b54-7969acd2dc93" (UID: "a56bac49-b398-4b61-9b54-7969acd2dc93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352348 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mfbr\" (UniqueName: \"kubernetes.io/projected/a56bac49-b398-4b61-9b54-7969acd2dc93-kube-api-access-5mfbr\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352380 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352390 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352398 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6ec9568-99c8-4bee-a97c-46400fcc0e73-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352407 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tggzq\" (UniqueName: \"kubernetes.io/projected/d6ec9568-99c8-4bee-a97c-46400fcc0e73-kube-api-access-tggzq\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.352415 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a56bac49-b398-4b61-9b54-7969acd2dc93-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.598871 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" event={"ID":"a56bac49-b398-4b61-9b54-7969acd2dc93","Type":"ContainerDied","Data":"cf4476f986138503b3408c91ec78e55f73e59536bc0804b03b95667a22e6c6a6"} Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.598939 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xh24t" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.598946 4733 scope.go:117] "RemoveContainer" containerID="9ab9c30c68bdbce8904477ffce48bb66e4703ed2c3ced83026789c4904ce7735" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.602572 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" event={"ID":"d6ec9568-99c8-4bee-a97c-46400fcc0e73","Type":"ContainerDied","Data":"642fd85cd38aa86a0c841253dfb5fda87ce7251b22474e12b3ef33923f92f9b2"} Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.602630 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tljb4" Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.640874 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.648368 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xh24t"] Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.656593 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:31:11 crc kubenswrapper[4733]: I0318 10:31:11.664637 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tljb4"] Mar 18 10:31:13 crc kubenswrapper[4733]: I0318 10:31:13.188936 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" path="/var/lib/kubelet/pods/a56bac49-b398-4b61-9b54-7969acd2dc93/volumes" Mar 18 10:31:13 crc kubenswrapper[4733]: I0318 10:31:13.190461 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" path="/var/lib/kubelet/pods/d6ec9568-99c8-4bee-a97c-46400fcc0e73/volumes" Mar 18 10:31:13 crc kubenswrapper[4733]: I0318 10:31:13.802447 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:15 crc kubenswrapper[4733]: I0318 10:31:15.374818 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:20 crc kubenswrapper[4733]: E0318 10:31:20.261966 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 18 10:31:20 crc kubenswrapper[4733]: E0318 10:31:20.262630 4733 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Mar 18 10:31:20 crc kubenswrapper[4733]: E0318 10:31:20.262819 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-77qsg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(55f0b230-09f2-4be2-aa1f-76a37f3fe30c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 10:31:20 crc kubenswrapper[4733]: E0318 10:31:20.264302 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="55f0b230-09f2-4be2-aa1f-76a37f3fe30c" Mar 18 10:31:20 crc kubenswrapper[4733]: I0318 10:31:20.680533 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerStarted","Data":"d334497d76d55a41bc85eeb60a10080fe3c2d76a582435f142bf759662b2098b"} Mar 18 10:31:20 crc kubenswrapper[4733]: E0318 10:31:20.682007 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="55f0b230-09f2-4be2-aa1f-76a37f3fe30c" Mar 18 10:31:20 crc kubenswrapper[4733]: I0318 10:31:20.861755 4733 scope.go:117] "RemoveContainer" containerID="0d7f9ebe26354b0fdcafbf9243319e234596728fe76f62846f8d6f2de9c01686" Mar 18 10:31:20 crc kubenswrapper[4733]: W0318 10:31:20.867388 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15df79ef_9d7a_4310_ba27_bdf8cb200f0f.slice/crio-1ad8ff69c6adc3dabc943b2d0fb235bc6e4c5c162e015c87081a85eb5257721c WatchSource:0}: Error finding container 1ad8ff69c6adc3dabc943b2d0fb235bc6e4c5c162e015c87081a85eb5257721c: Status 404 returned error can't find the container with id 1ad8ff69c6adc3dabc943b2d0fb235bc6e4c5c162e015c87081a85eb5257721c Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.079336 4733 scope.go:117] "RemoveContainer" containerID="aef65315ca9ffdfc8f038573ba4d7c7dad7e497bb3b3225251bd4117319d3146" Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.280502 4733 scope.go:117] "RemoveContainer" containerID="f1bfaec48682e00a29092241786f91d62461cdabac1600bdef26a808c7697bdd" Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.697483 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6trms" event={"ID":"e7849feb-5f1b-4b67-a3f7-8a419ebda0bd","Type":"ContainerStarted","Data":"89f64d45993ca9a528d13be0f8de5723e62397975bf7abe7b8b32795b7459d68"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.699784 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ljrgt" event={"ID":"d75a8d54-aca8-49cd-9062-6389baaf7a09","Type":"ContainerStarted","Data":"c9bad20081449289687cd76260cb1b4999faa861437220370b9811519c736d36"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.703026 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rh64b" event={"ID":"e3c842d3-b3dd-4cf2-9df0-16cea4061bc5","Type":"ContainerStarted","Data":"3b5a58b07f40c47d3810a2f753da5adaefb77778dab24aa461dc75931cdadf89"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.703571 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rh64b" Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.705115 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a8c27598-870d-4de0-a986-47042d7d6f4c","Type":"ContainerStarted","Data":"b65a0da116ca68b29800d866ab03f9d76e087c6073b691718aa6b8e2d620e06c"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.708817 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"dc60b49b-96fa-40fd-a8e5-40c810f5ef80","Type":"ContainerStarted","Data":"1996bad471830d839e4c51afbee66ae68a9d0380538f2fb4187d0e6cb1e23827"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.713381 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-6trms" podStartSLOduration=3.645966103 podStartE2EDuration="16.713368595s" podCreationTimestamp="2026-03-18 10:31:05 +0000 UTC" firstStartedPulling="2026-03-18 10:31:08.072163804 +0000 UTC m=+1107.563898129" lastFinishedPulling="2026-03-18 10:31:21.139566286 +0000 UTC m=+1120.631300621" observedRunningTime="2026-03-18 10:31:21.711143152 +0000 UTC m=+1121.202877467" watchObservedRunningTime="2026-03-18 10:31:21.713368595 +0000 UTC m=+1121.205102920" Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.721883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd66892e-808c-405a-ac8e-366b6ca8b148","Type":"ContainerStarted","Data":"123a82e05833cb782336a6b1c9d7f242be6afa92f95d5c9bcf898f5a656a5076"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.722603 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.725764 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerStarted","Data":"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.727071 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerStarted","Data":"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.727092 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerStarted","Data":"1ad8ff69c6adc3dabc943b2d0fb235bc6e4c5c162e015c87081a85eb5257721c"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.731684 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0868210e-9d93-4f63-b425-7db21f13cd90","Type":"ContainerStarted","Data":"32d5c065f6f149974bcb872025471f86b31a160e54f0c13567cc06fac5be118b"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.734549 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0208d826-df0f-41c8-83a7-821a21b7b85d","Type":"ContainerStarted","Data":"3874dea0990730a5c77d703a00ba1f3a9abfb0de15e4fe3acdd1a096b1bf4ecc"} Mar 18 10:31:21 crc kubenswrapper[4733]: I0318 10:31:21.793957 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rh64b" podStartSLOduration=5.761864006 podStartE2EDuration="21.793937545s" podCreationTimestamp="2026-03-18 10:31:00 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.032588958 +0000 UTC m=+1104.524323273" lastFinishedPulling="2026-03-18 10:31:21.064662487 +0000 UTC m=+1120.556396812" observedRunningTime="2026-03-18 10:31:21.785051443 +0000 UTC m=+1121.276785788" watchObservedRunningTime="2026-03-18 10:31:21.793937545 +0000 UTC m=+1121.285671870" Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.747270 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"1b521608cd076add0dc6ea82ec6fd5b69318ec8068de497c0a6615c97830553d"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.750496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a8c27598-870d-4de0-a986-47042d7d6f4c","Type":"ContainerStarted","Data":"07644ad53f49ea3b6fa47644d62cc985a5a2fc0647cac5021225d9225b5c3e56"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.753554 4733 generic.go:334] "Generic (PLEG): container finished" podID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerID="1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10" exitCode=0 Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.753632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerDied","Data":"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.753713 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerStarted","Data":"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.753805 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.762431 4733 generic.go:334] "Generic (PLEG): container finished" podID="d75a8d54-aca8-49cd-9062-6389baaf7a09" containerID="c9bad20081449289687cd76260cb1b4999faa861437220370b9811519c736d36" exitCode=0 Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.762577 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ljrgt" event={"ID":"d75a8d54-aca8-49cd-9062-6389baaf7a09","Type":"ContainerDied","Data":"c9bad20081449289687cd76260cb1b4999faa861437220370b9811519c736d36"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.770774 4733 generic.go:334] "Generic (PLEG): container finished" podID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerID="e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937" exitCode=0 Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.770916 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerDied","Data":"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.778664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0868210e-9d93-4f63-b425-7db21f13cd90","Type":"ContainerStarted","Data":"bf511622c4bebbf135ee7da9030a30c0e739ce4aa3cf29a5e9df7905992d8113"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.788459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"0fb5e774f72bc7530e7861681639d72697b8c0245883531528195b98bc45ea93"} Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.808529 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.435525729 podStartE2EDuration="27.808512206s" podCreationTimestamp="2026-03-18 10:30:55 +0000 UTC" firstStartedPulling="2026-03-18 10:31:04.663557394 +0000 UTC m=+1104.155291719" lastFinishedPulling="2026-03-18 10:31:13.036543871 +0000 UTC m=+1112.528278196" observedRunningTime="2026-03-18 10:31:21.874933677 +0000 UTC m=+1121.366668002" watchObservedRunningTime="2026-03-18 10:31:22.808512206 +0000 UTC m=+1122.300246551" Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.837902 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.675424847 podStartE2EDuration="19.837878907s" podCreationTimestamp="2026-03-18 10:31:03 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.941089007 +0000 UTC m=+1105.432823332" lastFinishedPulling="2026-03-18 10:31:21.103543037 +0000 UTC m=+1120.595277392" observedRunningTime="2026-03-18 10:31:22.831676162 +0000 UTC m=+1122.323410527" watchObservedRunningTime="2026-03-18 10:31:22.837878907 +0000 UTC m=+1122.329613232" Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.905171 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:22 crc kubenswrapper[4733]: I0318 10:31:22.944978 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.832178363 podStartE2EDuration="22.944959257s" podCreationTimestamp="2026-03-18 10:31:00 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.950124223 +0000 UTC m=+1105.441858548" lastFinishedPulling="2026-03-18 10:31:21.062905117 +0000 UTC m=+1120.554639442" observedRunningTime="2026-03-18 10:31:22.940558013 +0000 UTC m=+1122.432292338" watchObservedRunningTime="2026-03-18 10:31:22.944959257 +0000 UTC m=+1122.436693582" Mar 18 10:31:23 crc kubenswrapper[4733]: I0318 10:31:22.965461 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" podStartSLOduration=16.965438267 podStartE2EDuration="16.965438267s" podCreationTimestamp="2026-03-18 10:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:22.960300182 +0000 UTC m=+1122.452034517" watchObservedRunningTime="2026-03-18 10:31:22.965438267 +0000 UTC m=+1122.457172592" Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.905324 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.972963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerStarted","Data":"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b"} Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.974129 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.978676 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ljrgt" event={"ID":"d75a8d54-aca8-49cd-9062-6389baaf7a09","Type":"ContainerStarted","Data":"977df54e9e21058ae200793f039a61cc9e8ef7c488904477a179841b9c705157"} Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.978955 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.979090 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:24 crc kubenswrapper[4733]: I0318 10:31:24.979174 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ljrgt" event={"ID":"d75a8d54-aca8-49cd-9062-6389baaf7a09","Type":"ContainerStarted","Data":"0f25d0fd17f75d63c2160078393331e1696cc8d7b2ceed1f3ff8cb5a51395ea4"} Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.003129 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-rrvg6" podStartSLOduration=19.003108131 podStartE2EDuration="19.003108131s" podCreationTimestamp="2026-03-18 10:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:24.993691964 +0000 UTC m=+1124.485426289" watchObservedRunningTime="2026-03-18 10:31:25.003108131 +0000 UTC m=+1124.494842466" Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.028633 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-ljrgt" podStartSLOduration=15.157665566 podStartE2EDuration="25.028614703s" podCreationTimestamp="2026-03-18 10:31:00 +0000 UTC" firstStartedPulling="2026-03-18 10:31:04.728999556 +0000 UTC m=+1104.220733881" lastFinishedPulling="2026-03-18 10:31:14.599948693 +0000 UTC m=+1114.091683018" observedRunningTime="2026-03-18 10:31:25.026488642 +0000 UTC m=+1124.518222987" watchObservedRunningTime="2026-03-18 10:31:25.028614703 +0000 UTC m=+1124.520349038" Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.722078 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.778425 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.947407 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.987210 4733 generic.go:334] "Generic (PLEG): container finished" podID="dc60b49b-96fa-40fd-a8e5-40c810f5ef80" containerID="1996bad471830d839e4c51afbee66ae68a9d0380538f2fb4187d0e6cb1e23827" exitCode=0 Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.987251 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"dc60b49b-96fa-40fd-a8e5-40c810f5ef80","Type":"ContainerDied","Data":"1996bad471830d839e4c51afbee66ae68a9d0380538f2fb4187d0e6cb1e23827"} Mar 18 10:31:25 crc kubenswrapper[4733]: I0318 10:31:25.987901 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:26 crc kubenswrapper[4733]: I0318 10:31:26.781425 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 18 10:31:26 crc kubenswrapper[4733]: I0318 10:31:26.995531 4733 generic.go:334] "Generic (PLEG): container finished" podID="0208d826-df0f-41c8-83a7-821a21b7b85d" containerID="3874dea0990730a5c77d703a00ba1f3a9abfb0de15e4fe3acdd1a096b1bf4ecc" exitCode=0 Mar 18 10:31:26 crc kubenswrapper[4733]: I0318 10:31:26.995675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0208d826-df0f-41c8-83a7-821a21b7b85d","Type":"ContainerDied","Data":"3874dea0990730a5c77d703a00ba1f3a9abfb0de15e4fe3acdd1a096b1bf4ecc"} Mar 18 10:31:26 crc kubenswrapper[4733]: I0318 10:31:26.999659 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"dc60b49b-96fa-40fd-a8e5-40c810f5ef80","Type":"ContainerStarted","Data":"9ca944cb9aa23fc7d777229b0dcc77815b008c75e467a83230e086f840b8322e"} Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.056976 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.668398087 podStartE2EDuration="35.056957954s" podCreationTimestamp="2026-03-18 10:30:52 +0000 UTC" firstStartedPulling="2026-03-18 10:31:04.675094471 +0000 UTC m=+1104.166828796" lastFinishedPulling="2026-03-18 10:31:21.063654338 +0000 UTC m=+1120.555388663" observedRunningTime="2026-03-18 10:31:27.049637056 +0000 UTC m=+1126.541371381" watchObservedRunningTime="2026-03-18 10:31:27.056957954 +0000 UTC m=+1126.548692279" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.059007 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.234633 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 18 10:31:27 crc kubenswrapper[4733]: E0318 10:31:27.234943 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="init" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.234958 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="init" Mar 18 10:31:27 crc kubenswrapper[4733]: E0318 10:31:27.234996 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235002 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: E0318 10:31:27.235020 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="init" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235026 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="init" Mar 18 10:31:27 crc kubenswrapper[4733]: E0318 10:31:27.235038 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235045 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235179 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ec9568-99c8-4bee-a97c-46400fcc0e73" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235206 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a56bac49-b398-4b61-9b54-7969acd2dc93" containerName="dnsmasq-dns" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.235965 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.246813 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.246864 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-r2dgj" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.247244 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.247548 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.268908 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.276973 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-config\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-scripts\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277090 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdljd\" (UniqueName: \"kubernetes.io/projected/96c7007d-b722-4518-a298-269808d7dfc5-kube-api-access-qdljd\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/96c7007d-b722-4518-a298-269808d7dfc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.277215 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379269 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379352 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdljd\" (UniqueName: \"kubernetes.io/projected/96c7007d-b722-4518-a298-269808d7dfc5-kube-api-access-qdljd\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379402 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/96c7007d-b722-4518-a298-269808d7dfc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379505 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-config\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.379571 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-scripts\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.380536 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/96c7007d-b722-4518-a298-269808d7dfc5-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.380550 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-scripts\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.384985 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c7007d-b722-4518-a298-269808d7dfc5-config\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.385486 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.385962 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.389109 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c7007d-b722-4518-a298-269808d7dfc5-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.409751 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdljd\" (UniqueName: \"kubernetes.io/projected/96c7007d-b722-4518-a298-269808d7dfc5-kube-api-access-qdljd\") pod \"ovn-northd-0\" (UID: \"96c7007d-b722-4518-a298-269808d7dfc5\") " pod="openstack/ovn-northd-0" Mar 18 10:31:27 crc kubenswrapper[4733]: I0318 10:31:27.573939 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 18 10:31:28 crc kubenswrapper[4733]: I0318 10:31:28.007760 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0208d826-df0f-41c8-83a7-821a21b7b85d","Type":"ContainerStarted","Data":"4a28505a80aefa21b75c5f6a19c883fa6aad86f7eabe19eaca8f22fb40c92bd2"} Mar 18 10:31:28 crc kubenswrapper[4733]: I0318 10:31:28.031981 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 18 10:31:28 crc kubenswrapper[4733]: W0318 10:31:28.041605 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96c7007d_b722_4518_a298_269808d7dfc5.slice/crio-2a35526730e3ef6d6132defbecd56a9e4135c02443cefa8726bb9a19cdde7b0f WatchSource:0}: Error finding container 2a35526730e3ef6d6132defbecd56a9e4135c02443cefa8726bb9a19cdde7b0f: Status 404 returned error can't find the container with id 2a35526730e3ef6d6132defbecd56a9e4135c02443cefa8726bb9a19cdde7b0f Mar 18 10:31:28 crc kubenswrapper[4733]: I0318 10:31:28.052383 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.958102094 podStartE2EDuration="35.052354942s" podCreationTimestamp="2026-03-18 10:30:53 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.010963356 +0000 UTC m=+1104.502697671" lastFinishedPulling="2026-03-18 10:31:21.105216194 +0000 UTC m=+1120.596950519" observedRunningTime="2026-03-18 10:31:28.043287975 +0000 UTC m=+1127.535022340" watchObservedRunningTime="2026-03-18 10:31:28.052354942 +0000 UTC m=+1127.544089287" Mar 18 10:31:29 crc kubenswrapper[4733]: I0318 10:31:29.019896 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"96c7007d-b722-4518-a298-269808d7dfc5","Type":"ContainerStarted","Data":"2a35526730e3ef6d6132defbecd56a9e4135c02443cefa8726bb9a19cdde7b0f"} Mar 18 10:31:30 crc kubenswrapper[4733]: I0318 10:31:30.027387 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"96c7007d-b722-4518-a298-269808d7dfc5","Type":"ContainerStarted","Data":"bee4a30e65cbe318a5ccae86e3a925ae69bbb7b94faa9f5ce7d933a6bbfbce90"} Mar 18 10:31:30 crc kubenswrapper[4733]: I0318 10:31:30.027807 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 18 10:31:30 crc kubenswrapper[4733]: I0318 10:31:30.027818 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"96c7007d-b722-4518-a298-269808d7dfc5","Type":"ContainerStarted","Data":"67279349151b2537441da5a22cefbccf06053b41e8df222a62d51688bd4048e5"} Mar 18 10:31:30 crc kubenswrapper[4733]: I0318 10:31:30.045598 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.96362579 podStartE2EDuration="3.045580018s" podCreationTimestamp="2026-03-18 10:31:27 +0000 UTC" firstStartedPulling="2026-03-18 10:31:28.044587272 +0000 UTC m=+1127.536321607" lastFinishedPulling="2026-03-18 10:31:29.12654151 +0000 UTC m=+1128.618275835" observedRunningTime="2026-03-18 10:31:30.042668546 +0000 UTC m=+1129.534402881" watchObservedRunningTime="2026-03-18 10:31:30.045580018 +0000 UTC m=+1129.537314343" Mar 18 10:31:30 crc kubenswrapper[4733]: I0318 10:31:30.458438 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 18 10:31:31 crc kubenswrapper[4733]: I0318 10:31:31.598420 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:31 crc kubenswrapper[4733]: I0318 10:31:31.670375 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:31 crc kubenswrapper[4733]: I0318 10:31:31.743986 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.040072 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="dnsmasq-dns" containerID="cri-o://58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f" gracePeriod=10 Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.499881 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.585493 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc\") pod \"19376969-b236-4b21-b57f-3833a3c0c7b4\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.585584 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb\") pod \"19376969-b236-4b21-b57f-3833a3c0c7b4\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.585624 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config\") pod \"19376969-b236-4b21-b57f-3833a3c0c7b4\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.585841 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrzpq\" (UniqueName: \"kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq\") pod \"19376969-b236-4b21-b57f-3833a3c0c7b4\" (UID: \"19376969-b236-4b21-b57f-3833a3c0c7b4\") " Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.598633 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq" (OuterVolumeSpecName: "kube-api-access-nrzpq") pod "19376969-b236-4b21-b57f-3833a3c0c7b4" (UID: "19376969-b236-4b21-b57f-3833a3c0c7b4"). InnerVolumeSpecName "kube-api-access-nrzpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.625020 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "19376969-b236-4b21-b57f-3833a3c0c7b4" (UID: "19376969-b236-4b21-b57f-3833a3c0c7b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.631880 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19376969-b236-4b21-b57f-3833a3c0c7b4" (UID: "19376969-b236-4b21-b57f-3833a3c0c7b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.637088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config" (OuterVolumeSpecName: "config") pod "19376969-b236-4b21-b57f-3833a3c0c7b4" (UID: "19376969-b236-4b21-b57f-3833a3c0c7b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.687068 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrzpq\" (UniqueName: \"kubernetes.io/projected/19376969-b236-4b21-b57f-3833a3c0c7b4-kube-api-access-nrzpq\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.687101 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.687115 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:32 crc kubenswrapper[4733]: I0318 10:31:32.687127 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19376969-b236-4b21-b57f-3833a3c0c7b4-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.049910 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55f0b230-09f2-4be2-aa1f-76a37f3fe30c","Type":"ContainerStarted","Data":"416e5a0f591185bf6f079ff41418fd4332b9d5c05db84cf23ab9ab814c1c773f"} Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.050477 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.052408 4733 generic.go:334] "Generic (PLEG): container finished" podID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerID="58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f" exitCode=0 Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.052447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerDied","Data":"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f"} Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.052470 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" event={"ID":"19376969-b236-4b21-b57f-3833a3c0c7b4","Type":"ContainerDied","Data":"d334497d76d55a41bc85eeb60a10080fe3c2d76a582435f142bf759662b2098b"} Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.052490 4733 scope.go:117] "RemoveContainer" containerID="58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.052535 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-z55fc" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.091461 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.164764908 podStartE2EDuration="36.091431752s" podCreationTimestamp="2026-03-18 10:30:57 +0000 UTC" firstStartedPulling="2026-03-18 10:31:04.671965172 +0000 UTC m=+1104.163699497" lastFinishedPulling="2026-03-18 10:31:32.598632006 +0000 UTC m=+1132.090366341" observedRunningTime="2026-03-18 10:31:33.068652557 +0000 UTC m=+1132.560386882" watchObservedRunningTime="2026-03-18 10:31:33.091431752 +0000 UTC m=+1132.583166097" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.097026 4733 scope.go:117] "RemoveContainer" containerID="1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.107744 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.113782 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-z55fc"] Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.118734 4733 scope.go:117] "RemoveContainer" containerID="58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f" Mar 18 10:31:33 crc kubenswrapper[4733]: E0318 10:31:33.119175 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f\": container with ID starting with 58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f not found: ID does not exist" containerID="58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.119220 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f"} err="failed to get container status \"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f\": rpc error: code = NotFound desc = could not find container \"58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f\": container with ID starting with 58c76ce9b93e82f88df6f0ff6ee6c5e0a60fd6945aebd859567ed2120900425f not found: ID does not exist" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.119240 4733 scope.go:117] "RemoveContainer" containerID="1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10" Mar 18 10:31:33 crc kubenswrapper[4733]: E0318 10:31:33.119426 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10\": container with ID starting with 1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10 not found: ID does not exist" containerID="1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.119447 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10"} err="failed to get container status \"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10\": rpc error: code = NotFound desc = could not find container \"1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10\": container with ID starting with 1f57d62929ee7549431c614848f4b7ee032f7791e41c871d302148faa6989a10 not found: ID does not exist" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.190450 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" path="/var/lib/kubelet/pods/19376969-b236-4b21-b57f-3833a3c0c7b4/volumes" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.908176 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 18 10:31:33 crc kubenswrapper[4733]: I0318 10:31:33.908269 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 18 10:31:34 crc kubenswrapper[4733]: I0318 10:31:34.060714 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 18 10:31:34 crc kubenswrapper[4733]: I0318 10:31:34.145047 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 18 10:31:35 crc kubenswrapper[4733]: I0318 10:31:35.320835 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 18 10:31:35 crc kubenswrapper[4733]: I0318 10:31:35.320901 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 18 10:31:35 crc kubenswrapper[4733]: I0318 10:31:35.395318 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.160259 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.561755 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5e1e-account-create-update-r9bb4"] Mar 18 10:31:36 crc kubenswrapper[4733]: E0318 10:31:36.562132 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="init" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.562150 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="init" Mar 18 10:31:36 crc kubenswrapper[4733]: E0318 10:31:36.562211 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="dnsmasq-dns" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.562219 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="dnsmasq-dns" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.562425 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="19376969-b236-4b21-b57f-3833a3c0c7b4" containerName="dnsmasq-dns" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.563013 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.566085 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.579491 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5e1e-account-create-update-r9bb4"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.634329 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mr65v"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.635253 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.662349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.662472 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k7n9\" (UniqueName: \"kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.670480 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mr65v"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.732785 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5gwmb"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.733811 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.745379 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5gwmb"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.764039 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2qhr\" (UniqueName: \"kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.764732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.764805 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k7n9\" (UniqueName: \"kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.764855 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.765698 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.790609 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k7n9\" (UniqueName: \"kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9\") pod \"keystone-5e1e-account-create-update-r9bb4\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.838163 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0937-account-create-update-bfx7n"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.839372 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.841676 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.846139 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0937-account-create-update-bfx7n"] Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.867283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.867334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2qhr\" (UniqueName: \"kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.867396 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxpg\" (UniqueName: \"kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.867433 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.868310 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.880282 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:36 crc kubenswrapper[4733]: I0318 10:31:36.883756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2qhr\" (UniqueName: \"kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr\") pod \"keystone-db-create-mr65v\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.968519 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.968916 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.968994 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.969079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxpg\" (UniqueName: \"kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.969101 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrskv\" (UniqueName: \"kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:36.969939 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.042554 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxpg\" (UniqueName: \"kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg\") pod \"placement-db-create-5gwmb\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.050221 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.070674 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrskv\" (UniqueName: \"kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.071732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.071888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.090955 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrskv\" (UniqueName: \"kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv\") pod \"placement-0937-account-create-update-bfx7n\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.159863 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.367814 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5e1e-account-create-update-r9bb4"] Mar 18 10:31:37 crc kubenswrapper[4733]: W0318 10:31:37.372112 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6346115_9e7a_4489_916d_a129aa83a6dd.slice/crio-6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259 WatchSource:0}: Error finding container 6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259: Status 404 returned error can't find the container with id 6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259 Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.483774 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mr65v"] Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.569583 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5gwmb"] Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.663217 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0937-account-create-update-bfx7n"] Mar 18 10:31:37 crc kubenswrapper[4733]: W0318 10:31:37.684782 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07730b47_54ba_4b79_952e_6fb12b3b5279.slice/crio-4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84 WatchSource:0}: Error finding container 4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84: Status 404 returned error can't find the container with id 4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84 Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.802591 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.859547 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.860844 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.894455 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.985412 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.985577 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clgwm\" (UniqueName: \"kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.985647 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.985689 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:37 crc kubenswrapper[4733]: I0318 10:31:37.985754 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.091758 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clgwm\" (UniqueName: \"kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.091850 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.091897 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.091927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.091966 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.092906 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.093860 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.100251 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.112066 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.113283 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e1e-account-create-update-r9bb4" event={"ID":"a6346115-9e7a-4489-916d-a129aa83a6dd","Type":"ContainerStarted","Data":"1a52840f130018d9dd9a4d4957090d0bfe7cddccea8c86d998fc7ce63f88d2c3"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.113331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e1e-account-create-update-r9bb4" event={"ID":"a6346115-9e7a-4489-916d-a129aa83a6dd","Type":"ContainerStarted","Data":"6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.114319 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clgwm\" (UniqueName: \"kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm\") pod \"dnsmasq-dns-b8fbc5445-h4pnt\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.124692 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mr65v" event={"ID":"30a7c351-0be1-4547-bacc-8ff02cb59328","Type":"ContainerStarted","Data":"cb5c331f367d49d9d35cab0a581b0fd4e3d8921934861b35f887d6648ae09cfb"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.124757 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mr65v" event={"ID":"30a7c351-0be1-4547-bacc-8ff02cb59328","Type":"ContainerStarted","Data":"e708787994921982a80be3d54b5684318bbda7cf4ce148559d792c7f967c93d8"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.128004 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5gwmb" event={"ID":"0a5425fb-7059-4262-9c68-1420a5f3b4f1","Type":"ContainerStarted","Data":"ce9a99c6df86d54aacd4034e75a79275a1f1a3fe6a26a1b9d309967e3b0b146b"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.128052 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5gwmb" event={"ID":"0a5425fb-7059-4262-9c68-1420a5f3b4f1","Type":"ContainerStarted","Data":"f99fa9061248dffa81f529bc9a8356e3700d233c627489528d2203d1084f078a"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.138680 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0937-account-create-update-bfx7n" event={"ID":"07730b47-54ba-4b79-952e-6fb12b3b5279","Type":"ContainerStarted","Data":"881999bdad04a088176edfb2a1165638bbb818ce5892ed189c2612e4735ca703"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.138866 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5e1e-account-create-update-r9bb4" podStartSLOduration=2.138845979 podStartE2EDuration="2.138845979s" podCreationTimestamp="2026-03-18 10:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:38.128360592 +0000 UTC m=+1137.620094917" watchObservedRunningTime="2026-03-18 10:31:38.138845979 +0000 UTC m=+1137.630580304" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.138889 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0937-account-create-update-bfx7n" event={"ID":"07730b47-54ba-4b79-952e-6fb12b3b5279","Type":"ContainerStarted","Data":"4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84"} Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.156388 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-mr65v" podStartSLOduration=2.156372835 podStartE2EDuration="2.156372835s" podCreationTimestamp="2026-03-18 10:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:38.149753788 +0000 UTC m=+1137.641488123" watchObservedRunningTime="2026-03-18 10:31:38.156372835 +0000 UTC m=+1137.648107160" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.175870 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5gwmb" podStartSLOduration=2.175856936 podStartE2EDuration="2.175856936s" podCreationTimestamp="2026-03-18 10:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:38.168594071 +0000 UTC m=+1137.660328396" watchObservedRunningTime="2026-03-18 10:31:38.175856936 +0000 UTC m=+1137.667591261" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.205765 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.696573 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-0937-account-create-update-bfx7n" podStartSLOduration=2.696553071 podStartE2EDuration="2.696553071s" podCreationTimestamp="2026-03-18 10:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:38.185469028 +0000 UTC m=+1137.677203353" watchObservedRunningTime="2026-03-18 10:31:38.696553071 +0000 UTC m=+1138.188287396" Mar 18 10:31:38 crc kubenswrapper[4733]: I0318 10:31:38.699825 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:31:38 crc kubenswrapper[4733]: W0318 10:31:38.705836 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bdf8dbb_ffe1_48d1_9c79_22e37dd882be.slice/crio-5dd7dc77696d4097c1648883d1fba422fc00eb1a9ede4031a68c1b0d6e1e9d1c WatchSource:0}: Error finding container 5dd7dc77696d4097c1648883d1fba422fc00eb1a9ede4031a68c1b0d6e1e9d1c: Status 404 returned error can't find the container with id 5dd7dc77696d4097c1648883d1fba422fc00eb1a9ede4031a68c1b0d6e1e9d1c Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.077555 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.082517 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.085504 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.085706 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.085853 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.086001 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-bhzrc" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.113076 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.150766 4733 generic.go:334] "Generic (PLEG): container finished" podID="a6346115-9e7a-4489-916d-a129aa83a6dd" containerID="1a52840f130018d9dd9a4d4957090d0bfe7cddccea8c86d998fc7ce63f88d2c3" exitCode=0 Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.151162 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e1e-account-create-update-r9bb4" event={"ID":"a6346115-9e7a-4489-916d-a129aa83a6dd","Type":"ContainerDied","Data":"1a52840f130018d9dd9a4d4957090d0bfe7cddccea8c86d998fc7ce63f88d2c3"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.153241 4733 generic.go:334] "Generic (PLEG): container finished" podID="0a5425fb-7059-4262-9c68-1420a5f3b4f1" containerID="ce9a99c6df86d54aacd4034e75a79275a1f1a3fe6a26a1b9d309967e3b0b146b" exitCode=0 Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.153325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5gwmb" event={"ID":"0a5425fb-7059-4262-9c68-1420a5f3b4f1","Type":"ContainerDied","Data":"ce9a99c6df86d54aacd4034e75a79275a1f1a3fe6a26a1b9d309967e3b0b146b"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.155291 4733 generic.go:334] "Generic (PLEG): container finished" podID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerID="be1323a707d76c996153e9edb3286a8842293d7d0852b41ecba2e5d11f48e074" exitCode=0 Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.155353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" event={"ID":"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be","Type":"ContainerDied","Data":"be1323a707d76c996153e9edb3286a8842293d7d0852b41ecba2e5d11f48e074"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.155375 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" event={"ID":"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be","Type":"ContainerStarted","Data":"5dd7dc77696d4097c1648883d1fba422fc00eb1a9ede4031a68c1b0d6e1e9d1c"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.159086 4733 generic.go:334] "Generic (PLEG): container finished" podID="07730b47-54ba-4b79-952e-6fb12b3b5279" containerID="881999bdad04a088176edfb2a1165638bbb818ce5892ed189c2612e4735ca703" exitCode=0 Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.159198 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0937-account-create-update-bfx7n" event={"ID":"07730b47-54ba-4b79-952e-6fb12b3b5279","Type":"ContainerDied","Data":"881999bdad04a088176edfb2a1165638bbb818ce5892ed189c2612e4735ca703"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.160903 4733 generic.go:334] "Generic (PLEG): container finished" podID="30a7c351-0be1-4547-bacc-8ff02cb59328" containerID="cb5c331f367d49d9d35cab0a581b0fd4e3d8921934861b35f887d6648ae09cfb" exitCode=0 Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.160994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mr65v" event={"ID":"30a7c351-0be1-4547-bacc-8ff02cb59328","Type":"ContainerDied","Data":"cb5c331f367d49d9d35cab0a581b0fd4e3d8921934861b35f887d6648ae09cfb"} Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209156 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209349 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9pgx\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-kube-api-access-n9pgx\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209391 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-cache\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.209413 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-lock\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311259 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311346 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9pgx\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-kube-api-access-n9pgx\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-cache\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311442 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-lock\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311521 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.311526 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.311554 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.311612 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift podName:4f94cfc9-67cc-474c-8d99-58a9d4e0273f nodeName:}" failed. No retries permitted until 2026-03-18 10:31:39.811591225 +0000 UTC m=+1139.303325650 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift") pod "swift-storage-0" (UID: "4f94cfc9-67cc-474c-8d99-58a9d4e0273f") : configmap "swift-ring-files" not found Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.311728 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.312006 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-lock\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.312299 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-cache\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.312517 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.322130 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.331466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9pgx\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-kube-api-access-n9pgx\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.341126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: I0318 10:31:39.820150 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.820388 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.821249 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 18 10:31:39 crc kubenswrapper[4733]: E0318 10:31:39.821337 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift podName:4f94cfc9-67cc-474c-8d99-58a9d4e0273f nodeName:}" failed. No retries permitted until 2026-03-18 10:31:40.82131139 +0000 UTC m=+1140.313045725 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift") pod "swift-storage-0" (UID: "4f94cfc9-67cc-474c-8d99-58a9d4e0273f") : configmap "swift-ring-files" not found Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.172227 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" event={"ID":"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be","Type":"ContainerStarted","Data":"76c24ca485c5cd0e612df85bd0c8ef951256abf933fe1359b894cd82b8ea15fb"} Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.205635 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" podStartSLOduration=3.205619546 podStartE2EDuration="3.205619546s" podCreationTimestamp="2026-03-18 10:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:31:40.198578337 +0000 UTC m=+1139.690312672" watchObservedRunningTime="2026-03-18 10:31:40.205619546 +0000 UTC m=+1139.697353871" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.636521 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-fvlqt"] Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.637857 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.643492 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fvlqt"] Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.710101 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.715000 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5795-account-create-update-nkww7"] Mar 18 10:31:40 crc kubenswrapper[4733]: E0318 10:31:40.715421 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07730b47-54ba-4b79-952e-6fb12b3b5279" containerName="mariadb-account-create-update" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.715439 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="07730b47-54ba-4b79-952e-6fb12b3b5279" containerName="mariadb-account-create-update" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.715629 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="07730b47-54ba-4b79-952e-6fb12b3b5279" containerName="mariadb-account-create-update" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.716318 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.719129 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.743660 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqc4\" (UniqueName: \"kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.743695 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.753751 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5795-account-create-update-nkww7"] Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.786159 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.835776 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.841961 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.845709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts\") pod \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.845807 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts\") pod \"07730b47-54ba-4b79-952e-6fb12b3b5279\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.845836 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrskv\" (UniqueName: \"kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv\") pod \"07730b47-54ba-4b79-952e-6fb12b3b5279\" (UID: \"07730b47-54ba-4b79-952e-6fb12b3b5279\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.845878 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkxpg\" (UniqueName: \"kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg\") pod \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\" (UID: \"0a5425fb-7059-4262-9c68-1420a5f3b4f1\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846175 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a5425fb-7059-4262-9c68-1420a5f3b4f1" (UID: "0a5425fb-7059-4262-9c68-1420a5f3b4f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846210 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846255 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqc4\" (UniqueName: \"kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846308 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846353 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfscv\" (UniqueName: \"kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846462 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a5425fb-7059-4262-9c68-1420a5f3b4f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:40 crc kubenswrapper[4733]: E0318 10:31:40.846583 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 18 10:31:40 crc kubenswrapper[4733]: E0318 10:31:40.846640 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 18 10:31:40 crc kubenswrapper[4733]: E0318 10:31:40.846686 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift podName:4f94cfc9-67cc-474c-8d99-58a9d4e0273f nodeName:}" failed. No retries permitted until 2026-03-18 10:31:42.846668957 +0000 UTC m=+1142.338403282 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift") pod "swift-storage-0" (UID: "4f94cfc9-67cc-474c-8d99-58a9d4e0273f") : configmap "swift-ring-files" not found Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.846977 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.847648 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07730b47-54ba-4b79-952e-6fb12b3b5279" (UID: "07730b47-54ba-4b79-952e-6fb12b3b5279"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.866320 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv" (OuterVolumeSpecName: "kube-api-access-hrskv") pod "07730b47-54ba-4b79-952e-6fb12b3b5279" (UID: "07730b47-54ba-4b79-952e-6fb12b3b5279"). InnerVolumeSpecName "kube-api-access-hrskv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.867834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqc4\" (UniqueName: \"kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4\") pod \"glance-db-create-fvlqt\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.876495 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg" (OuterVolumeSpecName: "kube-api-access-vkxpg") pod "0a5425fb-7059-4262-9c68-1420a5f3b4f1" (UID: "0a5425fb-7059-4262-9c68-1420a5f3b4f1"). InnerVolumeSpecName "kube-api-access-vkxpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.954384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2qhr\" (UniqueName: \"kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr\") pod \"30a7c351-0be1-4547-bacc-8ff02cb59328\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.954441 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts\") pod \"a6346115-9e7a-4489-916d-a129aa83a6dd\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.954581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k7n9\" (UniqueName: \"kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9\") pod \"a6346115-9e7a-4489-916d-a129aa83a6dd\" (UID: \"a6346115-9e7a-4489-916d-a129aa83a6dd\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.954653 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts\") pod \"30a7c351-0be1-4547-bacc-8ff02cb59328\" (UID: \"30a7c351-0be1-4547-bacc-8ff02cb59328\") " Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.954876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.955027 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfscv\" (UniqueName: \"kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.955301 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07730b47-54ba-4b79-952e-6fb12b3b5279-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.955315 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrskv\" (UniqueName: \"kubernetes.io/projected/07730b47-54ba-4b79-952e-6fb12b3b5279-kube-api-access-hrskv\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.955328 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkxpg\" (UniqueName: \"kubernetes.io/projected/0a5425fb-7059-4262-9c68-1420a5f3b4f1-kube-api-access-vkxpg\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.956639 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6346115-9e7a-4489-916d-a129aa83a6dd" (UID: "a6346115-9e7a-4489-916d-a129aa83a6dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.956895 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30a7c351-0be1-4547-bacc-8ff02cb59328" (UID: "30a7c351-0be1-4547-bacc-8ff02cb59328"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.957493 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.959865 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9" (OuterVolumeSpecName: "kube-api-access-6k7n9") pod "a6346115-9e7a-4489-916d-a129aa83a6dd" (UID: "a6346115-9e7a-4489-916d-a129aa83a6dd"). InnerVolumeSpecName "kube-api-access-6k7n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.959913 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr" (OuterVolumeSpecName: "kube-api-access-w2qhr") pod "30a7c351-0be1-4547-bacc-8ff02cb59328" (UID: "30a7c351-0be1-4547-bacc-8ff02cb59328"). InnerVolumeSpecName "kube-api-access-w2qhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:40 crc kubenswrapper[4733]: I0318 10:31:40.972813 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfscv\" (UniqueName: \"kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv\") pod \"glance-5795-account-create-update-nkww7\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.056330 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k7n9\" (UniqueName: \"kubernetes.io/projected/a6346115-9e7a-4489-916d-a129aa83a6dd-kube-api-access-6k7n9\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.056364 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a7c351-0be1-4547-bacc-8ff02cb59328-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.056375 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2qhr\" (UniqueName: \"kubernetes.io/projected/30a7c351-0be1-4547-bacc-8ff02cb59328-kube-api-access-w2qhr\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.056386 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6346115-9e7a-4489-916d-a129aa83a6dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.083321 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.105171 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.200698 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e1e-account-create-update-r9bb4" event={"ID":"a6346115-9e7a-4489-916d-a129aa83a6dd","Type":"ContainerDied","Data":"6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259"} Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.200744 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6836746bdde3ee4414bdc9c36bf575e7d73bcbca50d7355ca4160c2ed4ad5259" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.200749 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e1e-account-create-update-r9bb4" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.202419 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mr65v" event={"ID":"30a7c351-0be1-4547-bacc-8ff02cb59328","Type":"ContainerDied","Data":"e708787994921982a80be3d54b5684318bbda7cf4ce148559d792c7f967c93d8"} Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.202448 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e708787994921982a80be3d54b5684318bbda7cf4ce148559d792c7f967c93d8" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.202508 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mr65v" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.209216 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5gwmb" event={"ID":"0a5425fb-7059-4262-9c68-1420a5f3b4f1","Type":"ContainerDied","Data":"f99fa9061248dffa81f529bc9a8356e3700d233c627489528d2203d1084f078a"} Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.209267 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99fa9061248dffa81f529bc9a8356e3700d233c627489528d2203d1084f078a" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.209356 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5gwmb" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.211884 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0937-account-create-update-bfx7n" event={"ID":"07730b47-54ba-4b79-952e-6fb12b3b5279","Type":"ContainerDied","Data":"4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84"} Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.211959 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dd3dd5968f22d9782c54516f96136d7407f843a2525378831477533afa22b84" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.211932 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0937-account-create-update-bfx7n" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.212068 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.707683 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5795-account-create-update-nkww7"] Mar 18 10:31:41 crc kubenswrapper[4733]: W0318 10:31:41.710515 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff5315db_fb68_4558_85c1_cf538d0e2770.slice/crio-2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47 WatchSource:0}: Error finding container 2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47: Status 404 returned error can't find the container with id 2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47 Mar 18 10:31:41 crc kubenswrapper[4733]: W0318 10:31:41.717999 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d4401f_2343_41fa_82ae_877674337bf4.slice/crio-b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde WatchSource:0}: Error finding container b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde: Status 404 returned error can't find the container with id b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.721857 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 18 10:31:41 crc kubenswrapper[4733]: I0318 10:31:41.725584 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fvlqt"] Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.220915 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5795-account-create-update-nkww7" event={"ID":"ff5315db-fb68-4558-85c1-cf538d0e2770","Type":"ContainerStarted","Data":"2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47"} Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.222261 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fvlqt" event={"ID":"84d4401f-2343-41fa-82ae-877674337bf4","Type":"ContainerStarted","Data":"b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde"} Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.548748 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-dj64z"] Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.549088 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6346115-9e7a-4489-916d-a129aa83a6dd" containerName="mariadb-account-create-update" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549105 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6346115-9e7a-4489-916d-a129aa83a6dd" containerName="mariadb-account-create-update" Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.549119 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a7c351-0be1-4547-bacc-8ff02cb59328" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549126 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a7c351-0be1-4547-bacc-8ff02cb59328" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.549136 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5425fb-7059-4262-9c68-1420a5f3b4f1" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549147 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5425fb-7059-4262-9c68-1420a5f3b4f1" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549328 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a7c351-0be1-4547-bacc-8ff02cb59328" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549342 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5425fb-7059-4262-9c68-1420a5f3b4f1" containerName="mariadb-database-create" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549356 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6346115-9e7a-4489-916d-a129aa83a6dd" containerName="mariadb-account-create-update" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.549845 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.553671 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.565095 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dj64z"] Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.695825 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x78zp\" (UniqueName: \"kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.695886 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.797952 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x78zp\" (UniqueName: \"kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.798023 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.798847 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.816853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x78zp\" (UniqueName: \"kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp\") pod \"root-account-create-update-dj64z\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.867855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.900161 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.900431 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.900448 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 18 10:31:42 crc kubenswrapper[4733]: E0318 10:31:42.900497 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift podName:4f94cfc9-67cc-474c-8d99-58a9d4e0273f nodeName:}" failed. No retries permitted until 2026-03-18 10:31:46.900479547 +0000 UTC m=+1146.392213872 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift") pod "swift-storage-0" (UID: "4f94cfc9-67cc-474c-8d99-58a9d4e0273f") : configmap "swift-ring-files" not found Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.909898 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-5ngrz"] Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.911107 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.916664 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.916676 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.926901 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.948308 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-5ngrz"] Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.976004 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nfmp2"] Mar 18 10:31:42 crc kubenswrapper[4733]: I0318 10:31:42.976984 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012520 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012631 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh78t\" (UniqueName: \"kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012807 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.012877 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.014969 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nfmp2"] Mar 18 10:31:43 crc kubenswrapper[4733]: E0318 10:31:43.018590 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-qh78t ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-5ngrz" podUID="ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.037412 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-5ngrz"] Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133320 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133386 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133419 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh78t\" (UniqueName: \"kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133481 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133504 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133536 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133586 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133629 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133645 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xzc\" (UniqueName: \"kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133662 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133681 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.133709 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.134803 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.135614 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.136070 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.142357 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.163789 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.167538 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.211076 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh78t\" (UniqueName: \"kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t\") pod \"swift-ring-rebalance-5ngrz\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272323 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xzc\" (UniqueName: \"kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272467 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272526 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.272587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.273373 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.274466 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.275744 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.280959 4733 generic.go:334] "Generic (PLEG): container finished" podID="84d4401f-2343-41fa-82ae-877674337bf4" containerID="c39830f7afc41d4e539449a18ea110efdcfaa942a6d99809d36a31233d6cb82b" exitCode=0 Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.281686 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.281808 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fvlqt" event={"ID":"84d4401f-2343-41fa-82ae-877674337bf4","Type":"ContainerDied","Data":"c39830f7afc41d4e539449a18ea110efdcfaa942a6d99809d36a31233d6cb82b"} Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.281831 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.284979 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.287522 4733 generic.go:334] "Generic (PLEG): container finished" podID="ff5315db-fb68-4558-85c1-cf538d0e2770" containerID="754b489534a1d2e07cfe28b803b1041b8c35b7a8d870ab7643873669d480405d" exitCode=0 Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.287592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.287963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5795-account-create-update-nkww7" event={"ID":"ff5315db-fb68-4558-85c1-cf538d0e2770","Type":"ContainerDied","Data":"754b489534a1d2e07cfe28b803b1041b8c35b7a8d870ab7643873669d480405d"} Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.304368 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xzc\" (UniqueName: \"kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc\") pod \"swift-ring-rebalance-nfmp2\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.311772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.337254 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-bhzrc" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.346261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.422073 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dj64z"] Mar 18 10:31:43 crc kubenswrapper[4733]: W0318 10:31:43.433719 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53658e2a_4376_49b7_82eb_f46c3dee3b6a.slice/crio-aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86 WatchSource:0}: Error finding container aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86: Status 404 returned error can't find the container with id aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86 Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.474771 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.474945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.475040 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.475139 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.481672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.481749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh78t\" (UniqueName: \"kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.482535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.482988 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.483053 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf\") pod \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\" (UID: \"ca2a461b-b7e7-4c74-9da8-2df85b95e6fd\") " Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.483620 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts" (OuterVolumeSpecName: "scripts") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.484737 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.484770 4733 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.484783 4733 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.488870 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.492881 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.505893 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.510753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t" (OuterVolumeSpecName: "kube-api-access-qh78t") pod "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" (UID: "ca2a461b-b7e7-4c74-9da8-2df85b95e6fd"). InnerVolumeSpecName "kube-api-access-qh78t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.590333 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh78t\" (UniqueName: \"kubernetes.io/projected/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-kube-api-access-qh78t\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.595582 4733 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.595703 4733 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.595777 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:43 crc kubenswrapper[4733]: W0318 10:31:43.789355 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e3fc960_7783_4952_90c9_1551c780ae03.slice/crio-9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45 WatchSource:0}: Error finding container 9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45: Status 404 returned error can't find the container with id 9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45 Mar 18 10:31:43 crc kubenswrapper[4733]: I0318 10:31:43.789683 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nfmp2"] Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.297392 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nfmp2" event={"ID":"5e3fc960-7783-4952-90c9-1551c780ae03","Type":"ContainerStarted","Data":"9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45"} Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.299910 4733 generic.go:334] "Generic (PLEG): container finished" podID="53658e2a-4376-49b7-82eb-f46c3dee3b6a" containerID="9bc7f39c4918c4a53f61ec2045418343aab6acedb5d7104271be60607764a8a9" exitCode=0 Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.299982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dj64z" event={"ID":"53658e2a-4376-49b7-82eb-f46c3dee3b6a","Type":"ContainerDied","Data":"9bc7f39c4918c4a53f61ec2045418343aab6acedb5d7104271be60607764a8a9"} Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.300015 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dj64z" event={"ID":"53658e2a-4376-49b7-82eb-f46c3dee3b6a","Type":"ContainerStarted","Data":"aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86"} Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.300036 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5ngrz" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.369177 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-5ngrz"] Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.375519 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-5ngrz"] Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.677704 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.784887 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.815649 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfscv\" (UniqueName: \"kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv\") pod \"ff5315db-fb68-4558-85c1-cf538d0e2770\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.815700 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts\") pod \"ff5315db-fb68-4558-85c1-cf538d0e2770\" (UID: \"ff5315db-fb68-4558-85c1-cf538d0e2770\") " Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.816557 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff5315db-fb68-4558-85c1-cf538d0e2770" (UID: "ff5315db-fb68-4558-85c1-cf538d0e2770"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.821489 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv" (OuterVolumeSpecName: "kube-api-access-jfscv") pod "ff5315db-fb68-4558-85c1-cf538d0e2770" (UID: "ff5315db-fb68-4558-85c1-cf538d0e2770"). InnerVolumeSpecName "kube-api-access-jfscv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.918078 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts\") pod \"84d4401f-2343-41fa-82ae-877674337bf4\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.918151 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqc4\" (UniqueName: \"kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4\") pod \"84d4401f-2343-41fa-82ae-877674337bf4\" (UID: \"84d4401f-2343-41fa-82ae-877674337bf4\") " Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.918697 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfscv\" (UniqueName: \"kubernetes.io/projected/ff5315db-fb68-4558-85c1-cf538d0e2770-kube-api-access-jfscv\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.918722 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5315db-fb68-4558-85c1-cf538d0e2770-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.919842 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84d4401f-2343-41fa-82ae-877674337bf4" (UID: "84d4401f-2343-41fa-82ae-877674337bf4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:44 crc kubenswrapper[4733]: I0318 10:31:44.922924 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4" (OuterVolumeSpecName: "kube-api-access-spqc4") pod "84d4401f-2343-41fa-82ae-877674337bf4" (UID: "84d4401f-2343-41fa-82ae-877674337bf4"). InnerVolumeSpecName "kube-api-access-spqc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.020443 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqc4\" (UniqueName: \"kubernetes.io/projected/84d4401f-2343-41fa-82ae-877674337bf4-kube-api-access-spqc4\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.020474 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d4401f-2343-41fa-82ae-877674337bf4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.189382 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2a461b-b7e7-4c74-9da8-2df85b95e6fd" path="/var/lib/kubelet/pods/ca2a461b-b7e7-4c74-9da8-2df85b95e6fd/volumes" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.309566 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fvlqt" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.309425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fvlqt" event={"ID":"84d4401f-2343-41fa-82ae-877674337bf4","Type":"ContainerDied","Data":"b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde"} Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.310338 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8b55af5d60c9bfa951021421dc142b3018bef82cf543260041378c3f0e7cbde" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.317396 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5795-account-create-update-nkww7" Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.317391 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5795-account-create-update-nkww7" event={"ID":"ff5315db-fb68-4558-85c1-cf538d0e2770","Type":"ContainerDied","Data":"2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47"} Mar 18 10:31:45 crc kubenswrapper[4733]: I0318 10:31:45.317443 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb7df9dc8f479fa452223731ed81c51e757e022bc1a6f323d9c6bcfdc951d47" Mar 18 10:31:46 crc kubenswrapper[4733]: I0318 10:31:46.947890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:46 crc kubenswrapper[4733]: E0318 10:31:46.948489 4733 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 18 10:31:46 crc kubenswrapper[4733]: E0318 10:31:46.948510 4733 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 18 10:31:46 crc kubenswrapper[4733]: E0318 10:31:46.948558 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift podName:4f94cfc9-67cc-474c-8d99-58a9d4e0273f nodeName:}" failed. No retries permitted until 2026-03-18 10:31:54.948542033 +0000 UTC m=+1154.440276368 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift") pod "swift-storage-0" (UID: "4f94cfc9-67cc-474c-8d99-58a9d4e0273f") : configmap "swift-ring-files" not found Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.334584 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dj64z" event={"ID":"53658e2a-4376-49b7-82eb-f46c3dee3b6a","Type":"ContainerDied","Data":"aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86"} Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.334628 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aebaf7db681fba4c59c36876ae242626d173f1734a9acd6bdfa4175925861c86" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.399139 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.558836 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts\") pod \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.559092 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x78zp\" (UniqueName: \"kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp\") pod \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\" (UID: \"53658e2a-4376-49b7-82eb-f46c3dee3b6a\") " Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.561487 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53658e2a-4376-49b7-82eb-f46c3dee3b6a" (UID: "53658e2a-4376-49b7-82eb-f46c3dee3b6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.563048 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp" (OuterVolumeSpecName: "kube-api-access-x78zp") pod "53658e2a-4376-49b7-82eb-f46c3dee3b6a" (UID: "53658e2a-4376-49b7-82eb-f46c3dee3b6a"). InnerVolumeSpecName "kube-api-access-x78zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.636749 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.661148 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53658e2a-4376-49b7-82eb-f46c3dee3b6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:47 crc kubenswrapper[4733]: I0318 10:31:47.661413 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x78zp\" (UniqueName: \"kubernetes.io/projected/53658e2a-4376-49b7-82eb-f46c3dee3b6a-kube-api-access-x78zp\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.207343 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.285931 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.286276 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-rrvg6" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="dnsmasq-dns" containerID="cri-o://e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b" gracePeriod=10 Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.344965 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dj64z" Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.344957 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nfmp2" event={"ID":"5e3fc960-7783-4952-90c9-1551c780ae03","Type":"ContainerStarted","Data":"dfd0abb25a1e6ce4147875c303d9c9787b741334508ecf0f7ab8f557701355ad"} Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.370646 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nfmp2" podStartSLOduration=2.888059724 podStartE2EDuration="6.370627827s" podCreationTimestamp="2026-03-18 10:31:42 +0000 UTC" firstStartedPulling="2026-03-18 10:31:43.796831733 +0000 UTC m=+1143.288566068" lastFinishedPulling="2026-03-18 10:31:47.279399846 +0000 UTC m=+1146.771134171" observedRunningTime="2026-03-18 10:31:48.362744623 +0000 UTC m=+1147.854478948" watchObservedRunningTime="2026-03-18 10:31:48.370627827 +0000 UTC m=+1147.862362152" Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.938956 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-dj64z"] Mar 18 10:31:48 crc kubenswrapper[4733]: I0318 10:31:48.952883 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-dj64z"] Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.199104 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53658e2a-4376-49b7-82eb-f46c3dee3b6a" path="/var/lib/kubelet/pods/53658e2a-4376-49b7-82eb-f46c3dee3b6a/volumes" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.247165 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.354891 4733 generic.go:334] "Generic (PLEG): container finished" podID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerID="e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b" exitCode=0 Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.354942 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rrvg6" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.354956 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerDied","Data":"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b"} Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.355000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rrvg6" event={"ID":"15df79ef-9d7a-4310-ba27-bdf8cb200f0f","Type":"ContainerDied","Data":"1ad8ff69c6adc3dabc943b2d0fb235bc6e4c5c162e015c87081a85eb5257721c"} Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.355023 4733 scope.go:117] "RemoveContainer" containerID="e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.402689 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb\") pod \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.402785 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb\") pod \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.402838 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc\") pod \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.402881 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5w85\" (UniqueName: \"kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85\") pod \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.402899 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config\") pod \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\" (UID: \"15df79ef-9d7a-4310-ba27-bdf8cb200f0f\") " Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.425741 4733 scope.go:117] "RemoveContainer" containerID="e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.437375 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85" (OuterVolumeSpecName: "kube-api-access-g5w85") pod "15df79ef-9d7a-4310-ba27-bdf8cb200f0f" (UID: "15df79ef-9d7a-4310-ba27-bdf8cb200f0f"). InnerVolumeSpecName "kube-api-access-g5w85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.448976 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15df79ef-9d7a-4310-ba27-bdf8cb200f0f" (UID: "15df79ef-9d7a-4310-ba27-bdf8cb200f0f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.462548 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15df79ef-9d7a-4310-ba27-bdf8cb200f0f" (UID: "15df79ef-9d7a-4310-ba27-bdf8cb200f0f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.479509 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config" (OuterVolumeSpecName: "config") pod "15df79ef-9d7a-4310-ba27-bdf8cb200f0f" (UID: "15df79ef-9d7a-4310-ba27-bdf8cb200f0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.492844 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15df79ef-9d7a-4310-ba27-bdf8cb200f0f" (UID: "15df79ef-9d7a-4310-ba27-bdf8cb200f0f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.504487 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5w85\" (UniqueName: \"kubernetes.io/projected/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-kube-api-access-g5w85\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.504927 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.504942 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.504952 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.504960 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15df79ef-9d7a-4310-ba27-bdf8cb200f0f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.541179 4733 scope.go:117] "RemoveContainer" containerID="e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b" Mar 18 10:31:49 crc kubenswrapper[4733]: E0318 10:31:49.541593 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b\": container with ID starting with e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b not found: ID does not exist" containerID="e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.541622 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b"} err="failed to get container status \"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b\": rpc error: code = NotFound desc = could not find container \"e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b\": container with ID starting with e9b6a3c12243c23c29491c07886c30384e8dc7b44b11048b3a29f67cf6a0e54b not found: ID does not exist" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.541644 4733 scope.go:117] "RemoveContainer" containerID="e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937" Mar 18 10:31:49 crc kubenswrapper[4733]: E0318 10:31:49.541935 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937\": container with ID starting with e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937 not found: ID does not exist" containerID="e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.542059 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937"} err="failed to get container status \"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937\": rpc error: code = NotFound desc = could not find container \"e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937\": container with ID starting with e32e3455bac4748d83432ff47c120fde28d910e682c598fcca6672f025864937 not found: ID does not exist" Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.686330 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:49 crc kubenswrapper[4733]: I0318 10:31:49.692059 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rrvg6"] Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.867572 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ptbmt"] Mar 18 10:31:50 crc kubenswrapper[4733]: E0318 10:31:50.868004 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d4401f-2343-41fa-82ae-877674337bf4" containerName="mariadb-database-create" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868022 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d4401f-2343-41fa-82ae-877674337bf4" containerName="mariadb-database-create" Mar 18 10:31:50 crc kubenswrapper[4733]: E0318 10:31:50.868030 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53658e2a-4376-49b7-82eb-f46c3dee3b6a" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868038 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="53658e2a-4376-49b7-82eb-f46c3dee3b6a" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: E0318 10:31:50.868063 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="dnsmasq-dns" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868069 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="dnsmasq-dns" Mar 18 10:31:50 crc kubenswrapper[4733]: E0318 10:31:50.868080 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5315db-fb68-4558-85c1-cf538d0e2770" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868086 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5315db-fb68-4558-85c1-cf538d0e2770" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: E0318 10:31:50.868098 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="init" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868104 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="init" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868297 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="53658e2a-4376-49b7-82eb-f46c3dee3b6a" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868316 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d4401f-2343-41fa-82ae-877674337bf4" containerName="mariadb-database-create" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868329 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5315db-fb68-4558-85c1-cf538d0e2770" containerName="mariadb-account-create-update" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868338 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" containerName="dnsmasq-dns" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.868895 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.871430 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7vljt" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.872006 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 18 10:31:50 crc kubenswrapper[4733]: I0318 10:31:50.886275 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ptbmt"] Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.029736 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.030199 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmmrf\" (UniqueName: \"kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.030295 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.030389 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.133090 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.133586 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.133814 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmmrf\" (UniqueName: \"kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.133890 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.139769 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.139906 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.151323 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.155957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmmrf\" (UniqueName: \"kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf\") pod \"glance-db-sync-ptbmt\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.183648 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ptbmt" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.190980 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15df79ef-9d7a-4310-ba27-bdf8cb200f0f" path="/var/lib/kubelet/pods/15df79ef-9d7a-4310-ba27-bdf8cb200f0f/volumes" Mar 18 10:31:51 crc kubenswrapper[4733]: I0318 10:31:51.773520 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ptbmt"] Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.387443 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ptbmt" event={"ID":"63c8f7bc-4162-4903-b3f9-96c8736a27b8","Type":"ContainerStarted","Data":"0e53d30c9ba54d0bab85e4e6730952c64da04afa3c77dd75422e1f34b5188d78"} Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.574865 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-gflvw"] Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.575879 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.579466 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.589958 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gflvw"] Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.670786 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.670876 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gqp6\" (UniqueName: \"kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.772818 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gqp6\" (UniqueName: \"kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.774453 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.773519 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.794701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gqp6\" (UniqueName: \"kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6\") pod \"root-account-create-update-gflvw\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:52 crc kubenswrapper[4733]: I0318 10:31:52.907668 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:53 crc kubenswrapper[4733]: I0318 10:31:53.401980 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gflvw"] Mar 18 10:31:53 crc kubenswrapper[4733]: W0318 10:31:53.421983 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4773e90c_6c0c_411c_810b_844d8570d4db.slice/crio-236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30 WatchSource:0}: Error finding container 236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30: Status 404 returned error can't find the container with id 236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30 Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.428661 4733 generic.go:334] "Generic (PLEG): container finished" podID="5e3fc960-7783-4952-90c9-1551c780ae03" containerID="dfd0abb25a1e6ce4147875c303d9c9787b741334508ecf0f7ab8f557701355ad" exitCode=0 Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.428957 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nfmp2" event={"ID":"5e3fc960-7783-4952-90c9-1551c780ae03","Type":"ContainerDied","Data":"dfd0abb25a1e6ce4147875c303d9c9787b741334508ecf0f7ab8f557701355ad"} Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.434320 4733 generic.go:334] "Generic (PLEG): container finished" podID="4773e90c-6c0c-411c-810b-844d8570d4db" containerID="fc8a98034f827fb8988cc2fa281e7a7c5e2bd32e772267e324591ed784c75b62" exitCode=0 Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.434425 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gflvw" event={"ID":"4773e90c-6c0c-411c-810b-844d8570d4db","Type":"ContainerDied","Data":"fc8a98034f827fb8988cc2fa281e7a7c5e2bd32e772267e324591ed784c75b62"} Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.434463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gflvw" event={"ID":"4773e90c-6c0c-411c-810b-844d8570d4db","Type":"ContainerStarted","Data":"236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30"} Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.436522 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="0fb5e774f72bc7530e7861681639d72697b8c0245883531528195b98bc45ea93" exitCode=0 Mar 18 10:31:54 crc kubenswrapper[4733]: I0318 10:31:54.436561 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"0fb5e774f72bc7530e7861681639d72697b8c0245883531528195b98bc45ea93"} Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.032979 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.053882 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4f94cfc9-67cc-474c-8d99-58a9d4e0273f-etc-swift\") pod \"swift-storage-0\" (UID: \"4f94cfc9-67cc-474c-8d99-58a9d4e0273f\") " pod="openstack/swift-storage-0" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.300531 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.445299 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="1b521608cd076add0dc6ea82ec6fd5b69318ec8068de497c0a6615c97830553d" exitCode=0 Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.445381 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"1b521608cd076add0dc6ea82ec6fd5b69318ec8068de497c0a6615c97830553d"} Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.455568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"001b9e894ecd9bef6f800f761255da30231b8fbfeb22f304bc145622a4998afa"} Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.456583 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.511467 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.509099846 podStartE2EDuration="1m4.511449354s" podCreationTimestamp="2026-03-18 10:30:51 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.062238827 +0000 UTC m=+1104.553973152" lastFinishedPulling="2026-03-18 10:31:21.064588335 +0000 UTC m=+1120.556322660" observedRunningTime="2026-03-18 10:31:55.509646563 +0000 UTC m=+1155.001380898" watchObservedRunningTime="2026-03-18 10:31:55.511449354 +0000 UTC m=+1155.003183679" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.881244 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rh64b" podUID="e3c842d3-b3dd-4cf2-9df0-16cea4061bc5" containerName="ovn-controller" probeResult="failure" output=< Mar 18 10:31:55 crc kubenswrapper[4733]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 18 10:31:55 crc kubenswrapper[4733]: > Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.888958 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.898556 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.899588 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.908218 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ljrgt" Mar 18 10:31:55 crc kubenswrapper[4733]: I0318 10:31:55.974652 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051209 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051322 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051406 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051453 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051491 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051597 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gqp6\" (UniqueName: \"kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6\") pod \"4773e90c-6c0c-411c-810b-844d8570d4db\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051625 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts\") pod \"4773e90c-6c0c-411c-810b-844d8570d4db\" (UID: \"4773e90c-6c0c-411c-810b-844d8570d4db\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.051670 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2xzc\" (UniqueName: \"kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc\") pod \"5e3fc960-7783-4952-90c9-1551c780ae03\" (UID: \"5e3fc960-7783-4952-90c9-1551c780ae03\") " Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.052941 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4773e90c-6c0c-411c-810b-844d8570d4db" (UID: "4773e90c-6c0c-411c-810b-844d8570d4db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.053902 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.055465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.056820 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc" (OuterVolumeSpecName: "kube-api-access-c2xzc") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "kube-api-access-c2xzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.058770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.059171 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6" (OuterVolumeSpecName: "kube-api-access-2gqp6") pod "4773e90c-6c0c-411c-810b-844d8570d4db" (UID: "4773e90c-6c0c-411c-810b-844d8570d4db"). InnerVolumeSpecName "kube-api-access-2gqp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.075562 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts" (OuterVolumeSpecName: "scripts") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.076945 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.080397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5e3fc960-7783-4952-90c9-1551c780ae03" (UID: "5e3fc960-7783-4952-90c9-1551c780ae03"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154173 4733 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154229 4733 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e3fc960-7783-4952-90c9-1551c780ae03-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154240 4733 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154252 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154264 4733 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e3fc960-7783-4952-90c9-1551c780ae03-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154276 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e3fc960-7783-4952-90c9-1551c780ae03-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154286 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gqp6\" (UniqueName: \"kubernetes.io/projected/4773e90c-6c0c-411c-810b-844d8570d4db-kube-api-access-2gqp6\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154301 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4773e90c-6c0c-411c-810b-844d8570d4db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.154312 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2xzc\" (UniqueName: \"kubernetes.io/projected/5e3fc960-7783-4952-90c9-1551c780ae03-kube-api-access-c2xzc\") on node \"crc\" DevicePath \"\"" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.169965 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rh64b-config-kckbs"] Mar 18 10:31:56 crc kubenswrapper[4733]: E0318 10:31:56.171471 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4773e90c-6c0c-411c-810b-844d8570d4db" containerName="mariadb-account-create-update" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.171500 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4773e90c-6c0c-411c-810b-844d8570d4db" containerName="mariadb-account-create-update" Mar 18 10:31:56 crc kubenswrapper[4733]: E0318 10:31:56.171535 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e3fc960-7783-4952-90c9-1551c780ae03" containerName="swift-ring-rebalance" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.171545 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e3fc960-7783-4952-90c9-1551c780ae03" containerName="swift-ring-rebalance" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.172723 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4773e90c-6c0c-411c-810b-844d8570d4db" containerName="mariadb-account-create-update" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.172759 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e3fc960-7783-4952-90c9-1551c780ae03" containerName="swift-ring-rebalance" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.174638 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.188507 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.209032 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rh64b-config-kckbs"] Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357033 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357095 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357536 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.357687 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlx9j\" (UniqueName: \"kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465549 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlx9j\" (UniqueName: \"kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465605 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465680 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465731 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465766 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465910 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465968 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.465998 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.466717 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.467972 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.468707 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nfmp2" event={"ID":"5e3fc960-7783-4952-90c9-1551c780ae03","Type":"ContainerDied","Data":"9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45"} Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.468736 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5077ef854e18aeb0823678f18007750be98c6ebb218cfdc6e156afe1f3ff45" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.468776 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nfmp2" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.470346 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gflvw" event={"ID":"4773e90c-6c0c-411c-810b-844d8570d4db","Type":"ContainerDied","Data":"236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30"} Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.470372 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="236e88250136b9b1d8c3c58d2a1f9c316bbac83718dd8135763f8a4a32d31a30" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.470383 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gflvw" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.471731 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"1082901e937aee3ff012135f915e36be2d201e2adb97cf8ab3cb6b5dbb1e9f6f"} Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.471886 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.473439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"4b22df5788ca868e21d692e0286a4fd86e269619d6c47e62147fd029667867ea"} Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.489549 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlx9j\" (UniqueName: \"kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j\") pod \"ovn-controller-rh64b-config-kckbs\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.499341 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:31:56 crc kubenswrapper[4733]: I0318 10:31:56.974848 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.428693781 podStartE2EDuration="1m6.974826055s" podCreationTimestamp="2026-03-18 10:30:50 +0000 UTC" firstStartedPulling="2026-03-18 10:31:05.052527152 +0000 UTC m=+1104.544261487" lastFinishedPulling="2026-03-18 10:31:14.598659436 +0000 UTC m=+1114.090393761" observedRunningTime="2026-03-18 10:31:56.500209345 +0000 UTC m=+1155.991943680" watchObservedRunningTime="2026-03-18 10:31:56.974826055 +0000 UTC m=+1156.466560380" Mar 18 10:31:57 crc kubenswrapper[4733]: I0318 10:31:57.022389 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rh64b-config-kckbs"] Mar 18 10:31:57 crc kubenswrapper[4733]: W0318 10:31:57.165528 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e628bba_84e9_4e1c_9e42_72277667b0a2.slice/crio-22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390 WatchSource:0}: Error finding container 22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390: Status 404 returned error can't find the container with id 22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390 Mar 18 10:31:57 crc kubenswrapper[4733]: I0318 10:31:57.484352 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rh64b-config-kckbs" event={"ID":"0e628bba-84e9-4e1c-9e42-72277667b0a2","Type":"ContainerStarted","Data":"22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.496832 4733 generic.go:334] "Generic (PLEG): container finished" podID="0e628bba-84e9-4e1c-9e42-72277667b0a2" containerID="6a586c5fd4b77aeae152ada8c17b0c5946fd162e927825804e11d81559ba17f0" exitCode=0 Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.496892 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rh64b-config-kckbs" event={"ID":"0e628bba-84e9-4e1c-9e42-72277667b0a2","Type":"ContainerDied","Data":"6a586c5fd4b77aeae152ada8c17b0c5946fd162e927825804e11d81559ba17f0"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.503468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"d19587f53ed713b0ff5fa1da4b4e7eba2645babc58913c28cba7a274f45c17df"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.503824 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"b4be968df0168e894e0715f49feb099d6e403dc44a8d1eee90a40ddfac216669"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.503840 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"dbfe977643a9bde3af88327de7777951f800a8fc1d72a36faddff72583073343"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.503852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"266fbf2493fb04519ff239ccf13b44e1a705933d786e0b3b1056deede6dccf8c"} Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.948953 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-gflvw"] Mar 18 10:31:58 crc kubenswrapper[4733]: I0318 10:31:58.955139 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-gflvw"] Mar 18 10:31:59 crc kubenswrapper[4733]: I0318 10:31:59.184821 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4773e90c-6c0c-411c-810b-844d8570d4db" path="/var/lib/kubelet/pods/4773e90c-6c0c-411c-810b-844d8570d4db/volumes" Mar 18 10:31:59 crc kubenswrapper[4733]: I0318 10:31:59.516138 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="001b9e894ecd9bef6f800f761255da30231b8fbfeb22f304bc145622a4998afa" exitCode=0 Mar 18 10:31:59 crc kubenswrapper[4733]: I0318 10:31:59.516217 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"001b9e894ecd9bef6f800f761255da30231b8fbfeb22f304bc145622a4998afa"} Mar 18 10:31:59 crc kubenswrapper[4733]: I0318 10:31:59.517826 4733 scope.go:117] "RemoveContainer" containerID="001b9e894ecd9bef6f800f761255da30231b8fbfeb22f304bc145622a4998afa" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.133827 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563832-njktw"] Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.134993 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.141366 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563832-njktw"] Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.171281 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.171499 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.171696 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.232365 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47d2\" (UniqueName: \"kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2\") pod \"auto-csr-approver-29563832-njktw\" (UID: \"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e\") " pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.334460 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47d2\" (UniqueName: \"kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2\") pod \"auto-csr-approver-29563832-njktw\" (UID: \"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e\") " pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.375384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47d2\" (UniqueName: \"kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2\") pod \"auto-csr-approver-29563832-njktw\" (UID: \"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e\") " pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.499790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.524497 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="1082901e937aee3ff012135f915e36be2d201e2adb97cf8ab3cb6b5dbb1e9f6f" exitCode=0 Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.524542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"1082901e937aee3ff012135f915e36be2d201e2adb97cf8ab3cb6b5dbb1e9f6f"} Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.525174 4733 scope.go:117] "RemoveContainer" containerID="1082901e937aee3ff012135f915e36be2d201e2adb97cf8ab3cb6b5dbb1e9f6f" Mar 18 10:32:00 crc kubenswrapper[4733]: I0318 10:32:00.876721 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rh64b" Mar 18 10:32:03 crc kubenswrapper[4733]: I0318 10:32:03.963860 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-xb87f"] Mar 18 10:32:03 crc kubenswrapper[4733]: I0318 10:32:03.965869 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:03 crc kubenswrapper[4733]: I0318 10:32:03.968002 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:03.972002 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xb87f"] Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.096008 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.096120 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87bq\" (UniqueName: \"kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.197097 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87bq\" (UniqueName: \"kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.197421 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.198157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.220029 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87bq\" (UniqueName: \"kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq\") pod \"root-account-create-update-xb87f\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:04 crc kubenswrapper[4733]: I0318 10:32:04.327351 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:08 crc kubenswrapper[4733]: E0318 10:32:08.423472 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Mar 18 10:32:08 crc kubenswrapper[4733]: E0318 10:32:08.424086 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vmmrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-ptbmt_openstack(63c8f7bc-4162-4903-b3f9-96c8736a27b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 10:32:08 crc kubenswrapper[4733]: E0318 10:32:08.425512 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-ptbmt" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.583137 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.613371 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rh64b-config-kckbs" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.613704 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rh64b-config-kckbs" event={"ID":"0e628bba-84e9-4e1c-9e42-72277667b0a2","Type":"ContainerDied","Data":"22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390"} Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.613733 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22ce6d62b2c247efc2f78a573eaad62efb486d3285646f6c9d81f05e03bd4390" Mar 18 10:32:08 crc kubenswrapper[4733]: E0318 10:32:08.615055 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-ptbmt" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.674943 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675041 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675085 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlx9j\" (UniqueName: \"kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675507 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run\") pod \"0e628bba-84e9-4e1c-9e42-72277667b0a2\" (UID: \"0e628bba-84e9-4e1c-9e42-72277667b0a2\") " Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675782 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.675832 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.676868 4733 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.677068 4733 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.676902 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts" (OuterVolumeSpecName: "scripts") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.676928 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run" (OuterVolumeSpecName: "var-run") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.676985 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.681322 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j" (OuterVolumeSpecName: "kube-api-access-mlx9j") pod "0e628bba-84e9-4e1c-9e42-72277667b0a2" (UID: "0e628bba-84e9-4e1c-9e42-72277667b0a2"). InnerVolumeSpecName "kube-api-access-mlx9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.778653 4733 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e628bba-84e9-4e1c-9e42-72277667b0a2-var-run\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.778686 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.778697 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlx9j\" (UniqueName: \"kubernetes.io/projected/0e628bba-84e9-4e1c-9e42-72277667b0a2-kube-api-access-mlx9j\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.778708 4733 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e628bba-84e9-4e1c-9e42-72277667b0a2-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.944253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563832-njktw"] Mar 18 10:32:08 crc kubenswrapper[4733]: W0318 10:32:08.947211 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac26b4cb_ac0b_4b78_9c5e_60c6563b478e.slice/crio-8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710 WatchSource:0}: Error finding container 8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710: Status 404 returned error can't find the container with id 8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710 Mar 18 10:32:08 crc kubenswrapper[4733]: I0318 10:32:08.994781 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xb87f"] Mar 18 10:32:09 crc kubenswrapper[4733]: W0318 10:32:09.162233 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd533f566_eded_44ca_b276_7e3d437f9fce.slice/crio-e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a WatchSource:0}: Error finding container e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a: Status 404 returned error can't find the container with id e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.639752 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.641299 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.649400 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.649815 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.665559 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"f09f54d8c78fa82ebedc2ac49c3d041989b31b6f8ff99c6e4662292723ec43a0"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.665610 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"9f55e786080c5ae38bb5df6db279e3cae26306d1b2cb1db2ee0263184a18fa92"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.669585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563832-njktw" event={"ID":"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e","Type":"ContainerStarted","Data":"8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.680620 4733 generic.go:334] "Generic (PLEG): container finished" podID="d533f566-eded-44ca-b276-7e3d437f9fce" containerID="7076d89bfeedd95679091846270edbde667d954c3e6fdb8ee00f499b50144915" exitCode=0 Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.680677 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xb87f" event={"ID":"d533f566-eded-44ca-b276-7e3d437f9fce","Type":"ContainerDied","Data":"7076d89bfeedd95679091846270edbde667d954c3e6fdb8ee00f499b50144915"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.680704 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xb87f" event={"ID":"d533f566-eded-44ca-b276-7e3d437f9fce","Type":"ContainerStarted","Data":"e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a"} Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.740283 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rh64b-config-kckbs"] Mar 18 10:32:09 crc kubenswrapper[4733]: I0318 10:32:09.748425 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rh64b-config-kckbs"] Mar 18 10:32:10 crc kubenswrapper[4733]: I0318 10:32:10.693845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"03b367733a02dfdbd7555ecf7e95eae5a15fd9b63d6399da7d29fc384dd98c9e"} Mar 18 10:32:10 crc kubenswrapper[4733]: I0318 10:32:10.694128 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"1109244a1ec38349bb9c33394e754f6ca373727120e2016826fdcbc9e7725e44"} Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.186827 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e628bba-84e9-4e1c-9e42-72277667b0a2" path="/var/lib/kubelet/pods/0e628bba-84e9-4e1c-9e42-72277667b0a2/volumes" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.275979 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.424911 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n87bq\" (UniqueName: \"kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq\") pod \"d533f566-eded-44ca-b276-7e3d437f9fce\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.425450 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts\") pod \"d533f566-eded-44ca-b276-7e3d437f9fce\" (UID: \"d533f566-eded-44ca-b276-7e3d437f9fce\") " Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.426060 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d533f566-eded-44ca-b276-7e3d437f9fce" (UID: "d533f566-eded-44ca-b276-7e3d437f9fce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.429418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq" (OuterVolumeSpecName: "kube-api-access-n87bq") pod "d533f566-eded-44ca-b276-7e3d437f9fce" (UID: "d533f566-eded-44ca-b276-7e3d437f9fce"). InnerVolumeSpecName "kube-api-access-n87bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.527581 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n87bq\" (UniqueName: \"kubernetes.io/projected/d533f566-eded-44ca-b276-7e3d437f9fce-kube-api-access-n87bq\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.527611 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d533f566-eded-44ca-b276-7e3d437f9fce-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.708111 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"44cc72cf61efaa659b87c1775077631473a06083fb71d0347c99e0d6c7c980e0"} Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.708153 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"b254459b4ffc3b00f73b9608b3610886dd56d5840666881fbaa05386bf9269fe"} Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.709822 4733 generic.go:334] "Generic (PLEG): container finished" podID="ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" containerID="da48f5028812280b5314f3d818c71b3049bdb0d8b1d5755bc74f1fedad4676d7" exitCode=0 Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.709922 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563832-njktw" event={"ID":"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e","Type":"ContainerDied","Data":"da48f5028812280b5314f3d818c71b3049bdb0d8b1d5755bc74f1fedad4676d7"} Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.710977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xb87f" event={"ID":"d533f566-eded-44ca-b276-7e3d437f9fce","Type":"ContainerDied","Data":"e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a"} Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.711004 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79e61d37c63bf1173443cab053dcac88ea174ce8aa0f6e02aa470300d87201a" Mar 18 10:32:11 crc kubenswrapper[4733]: I0318 10:32:11.711058 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xb87f" Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.726559 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"d9de382521acac6cb3603510e0e270022fc866b500e77b3ff7d004243afaca92"} Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.727883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"100e25c73401fec565df6ba0895b1f5ea995dfe6b674db5b47d91ed3fc50711d"} Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.728747 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"a147ff81dceb9facdc603abb3d83a7c1416cf9e85e710a7991b6a0626233c259"} Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.728850 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"6e01f54b2250bac6d78a8f11659c5103ad761b5a72b5739349656250527ed43e"} Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.728953 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4f94cfc9-67cc-474c-8d99-58a9d4e0273f","Type":"ContainerStarted","Data":"95ba0cc3f48d5f96dab0f77a9064545ae6f43f2040f94300ac8d0f8925102f4f"} Mar 18 10:32:12 crc kubenswrapper[4733]: I0318 10:32:12.768435 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.578083254 podStartE2EDuration="34.768414202s" podCreationTimestamp="2026-03-18 10:31:38 +0000 UTC" firstStartedPulling="2026-03-18 10:31:56.000950926 +0000 UTC m=+1155.492685251" lastFinishedPulling="2026-03-18 10:32:11.191281874 +0000 UTC m=+1170.683016199" observedRunningTime="2026-03-18 10:32:12.758109242 +0000 UTC m=+1172.249843567" watchObservedRunningTime="2026-03-18 10:32:12.768414202 +0000 UTC m=+1172.260148527" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.035996 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:13 crc kubenswrapper[4733]: E0318 10:32:13.036626 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e628bba-84e9-4e1c-9e42-72277667b0a2" containerName="ovn-config" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.036642 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e628bba-84e9-4e1c-9e42-72277667b0a2" containerName="ovn-config" Mar 18 10:32:13 crc kubenswrapper[4733]: E0318 10:32:13.036653 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d533f566-eded-44ca-b276-7e3d437f9fce" containerName="mariadb-account-create-update" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.036660 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d533f566-eded-44ca-b276-7e3d437f9fce" containerName="mariadb-account-create-update" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.036815 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e628bba-84e9-4e1c-9e42-72277667b0a2" containerName="ovn-config" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.036840 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d533f566-eded-44ca-b276-7e3d437f9fce" containerName="mariadb-account-create-update" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.037594 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.039248 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.075017 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.078379 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160591 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160670 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160692 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160738 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbm6m\" (UniqueName: \"kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160755 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.160785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.261950 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n47d2\" (UniqueName: \"kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2\") pod \"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e\" (UID: \"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e\") " Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262381 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbm6m\" (UniqueName: \"kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262416 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262469 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.262534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.264552 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.264562 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.264898 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.264900 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.267535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2" (OuterVolumeSpecName: "kube-api-access-n47d2") pod "ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" (UID: "ac26b4cb-ac0b-4b78-9c5e-60c6563b478e"). InnerVolumeSpecName "kube-api-access-n47d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.268116 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.279882 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbm6m\" (UniqueName: \"kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m\") pod \"dnsmasq-dns-5c79d794d7-5vkkr\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.364524 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n47d2\" (UniqueName: \"kubernetes.io/projected/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e-kube-api-access-n47d2\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.387808 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.737088 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563832-njktw" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.739964 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563832-njktw" event={"ID":"ac26b4cb-ac0b-4b78-9c5e-60c6563b478e","Type":"ContainerDied","Data":"8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710"} Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.740006 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac1b0f08ec5d849a8bed24cb2098e48f5ac74bb6bd23eef8fa08f76cffaa710" Mar 18 10:32:13 crc kubenswrapper[4733]: I0318 10:32:13.963050 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:13 crc kubenswrapper[4733]: W0318 10:32:13.974248 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0c771ba_dbb2_470b_b19c_8c8fdefbdd6d.slice/crio-e1e841a260dd28bf7f12085c583e93ce7e0418239f8aa3a3290f316b2610e14f WatchSource:0}: Error finding container e1e841a260dd28bf7f12085c583e93ce7e0418239f8aa3a3290f316b2610e14f: Status 404 returned error can't find the container with id e1e841a260dd28bf7f12085c583e93ce7e0418239f8aa3a3290f316b2610e14f Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.142298 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563826-tfzqx"] Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.150394 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563826-tfzqx"] Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.746079 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683" exitCode=0 Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.746140 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683"} Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.746202 4733 scope.go:117] "RemoveContainer" containerID="001b9e894ecd9bef6f800f761255da30231b8fbfeb22f304bc145622a4998afa" Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.746806 4733 scope.go:117] "RemoveContainer" containerID="9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683" Mar 18 10:32:14 crc kubenswrapper[4733]: E0318 10:32:14.747044 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 10s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.749596 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1" exitCode=0 Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.749668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1"} Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.750023 4733 scope.go:117] "RemoveContainer" containerID="f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1" Mar 18 10:32:14 crc kubenswrapper[4733]: E0318 10:32:14.750185 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 10s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.752884 4733 generic.go:334] "Generic (PLEG): container finished" podID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerID="d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535" exitCode=0 Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.752914 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" event={"ID":"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d","Type":"ContainerDied","Data":"d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535"} Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.752941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" event={"ID":"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d","Type":"ContainerStarted","Data":"e1e841a260dd28bf7f12085c583e93ce7e0418239f8aa3a3290f316b2610e14f"} Mar 18 10:32:14 crc kubenswrapper[4733]: I0318 10:32:14.906511 4733 scope.go:117] "RemoveContainer" containerID="1082901e937aee3ff012135f915e36be2d201e2adb97cf8ab3cb6b5dbb1e9f6f" Mar 18 10:32:15 crc kubenswrapper[4733]: I0318 10:32:15.191682 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9f28a6-6f4c-440b-abfc-cca26041cbef" path="/var/lib/kubelet/pods/eb9f28a6-6f4c-440b-abfc-cca26041cbef/volumes" Mar 18 10:32:15 crc kubenswrapper[4733]: I0318 10:32:15.782350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" event={"ID":"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d","Type":"ContainerStarted","Data":"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db"} Mar 18 10:32:15 crc kubenswrapper[4733]: I0318 10:32:15.782514 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:15 crc kubenswrapper[4733]: I0318 10:32:15.803828 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" podStartSLOduration=2.803803945 podStartE2EDuration="2.803803945s" podCreationTimestamp="2026-03-18 10:32:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:32:15.79761485 +0000 UTC m=+1175.289349215" watchObservedRunningTime="2026-03-18 10:32:15.803803945 +0000 UTC m=+1175.295538290" Mar 18 10:32:20 crc kubenswrapper[4733]: I0318 10:32:20.179293 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:32:21 crc kubenswrapper[4733]: I0318 10:32:21.858885 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ptbmt" event={"ID":"63c8f7bc-4162-4903-b3f9-96c8736a27b8","Type":"ContainerStarted","Data":"5eade19f29d1bfd378a24e80adf648a30d27b707a4c07763d5b7990ffd79ce55"} Mar 18 10:32:21 crc kubenswrapper[4733]: I0318 10:32:21.890769 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ptbmt" podStartSLOduration=3.034984458 podStartE2EDuration="31.890718105s" podCreationTimestamp="2026-03-18 10:31:50 +0000 UTC" firstStartedPulling="2026-03-18 10:31:51.77937263 +0000 UTC m=+1151.271106955" lastFinishedPulling="2026-03-18 10:32:20.635106257 +0000 UTC m=+1180.126840602" observedRunningTime="2026-03-18 10:32:21.88912331 +0000 UTC m=+1181.380857645" watchObservedRunningTime="2026-03-18 10:32:21.890718105 +0000 UTC m=+1181.382452420" Mar 18 10:32:23 crc kubenswrapper[4733]: I0318 10:32:23.389422 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:23 crc kubenswrapper[4733]: I0318 10:32:23.461086 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:32:23 crc kubenswrapper[4733]: I0318 10:32:23.461554 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="dnsmasq-dns" containerID="cri-o://76c24ca485c5cd0e612df85bd0c8ef951256abf933fe1359b894cd82b8ea15fb" gracePeriod=10 Mar 18 10:32:23 crc kubenswrapper[4733]: I0318 10:32:23.890815 4733 generic.go:334] "Generic (PLEG): container finished" podID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerID="76c24ca485c5cd0e612df85bd0c8ef951256abf933fe1359b894cd82b8ea15fb" exitCode=0 Mar 18 10:32:23 crc kubenswrapper[4733]: I0318 10:32:23.890854 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" event={"ID":"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be","Type":"ContainerDied","Data":"76c24ca485c5cd0e612df85bd0c8ef951256abf933fe1359b894cd82b8ea15fb"} Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.046082 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.143803 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc\") pod \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.143918 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb\") pod \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.143962 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config\") pod \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.144005 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb\") pod \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.144154 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clgwm\" (UniqueName: \"kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm\") pod \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\" (UID: \"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be\") " Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.157454 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm" (OuterVolumeSpecName: "kube-api-access-clgwm") pod "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" (UID: "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be"). InnerVolumeSpecName "kube-api-access-clgwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.208753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" (UID: "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.215179 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" (UID: "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.233383 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" (UID: "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.249663 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config" (OuterVolumeSpecName: "config") pod "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" (UID: "7bdf8dbb-ffe1-48d1-9c79-22e37dd882be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.251327 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.251353 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.251366 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.251379 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clgwm\" (UniqueName: \"kubernetes.io/projected/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-kube-api-access-clgwm\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.251401 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.902723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" event={"ID":"7bdf8dbb-ffe1-48d1-9c79-22e37dd882be","Type":"ContainerDied","Data":"5dd7dc77696d4097c1648883d1fba422fc00eb1a9ede4031a68c1b0d6e1e9d1c"} Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.902812 4733 scope.go:117] "RemoveContainer" containerID="76c24ca485c5cd0e612df85bd0c8ef951256abf933fe1359b894cd82b8ea15fb" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.902867 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-h4pnt" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.930105 4733 scope.go:117] "RemoveContainer" containerID="be1323a707d76c996153e9edb3286a8842293d7d0852b41ecba2e5d11f48e074" Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.966947 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:32:24 crc kubenswrapper[4733]: I0318 10:32:24.976099 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-h4pnt"] Mar 18 10:32:25 crc kubenswrapper[4733]: I0318 10:32:25.185861 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" path="/var/lib/kubelet/pods/7bdf8dbb-ffe1-48d1-9c79-22e37dd882be/volumes" Mar 18 10:32:26 crc kubenswrapper[4733]: I0318 10:32:26.176558 4733 scope.go:117] "RemoveContainer" containerID="f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1" Mar 18 10:32:26 crc kubenswrapper[4733]: I0318 10:32:26.920627 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4"} Mar 18 10:32:26 crc kubenswrapper[4733]: I0318 10:32:26.921155 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:32:27 crc kubenswrapper[4733]: I0318 10:32:27.931954 4733 generic.go:334] "Generic (PLEG): container finished" podID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" containerID="5eade19f29d1bfd378a24e80adf648a30d27b707a4c07763d5b7990ffd79ce55" exitCode=0 Mar 18 10:32:27 crc kubenswrapper[4733]: I0318 10:32:27.932292 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ptbmt" event={"ID":"63c8f7bc-4162-4903-b3f9-96c8736a27b8","Type":"ContainerDied","Data":"5eade19f29d1bfd378a24e80adf648a30d27b707a4c07763d5b7990ffd79ce55"} Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.176317 4733 scope.go:117] "RemoveContainer" containerID="9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.341106 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ptbmt" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.433065 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmmrf\" (UniqueName: \"kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf\") pod \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.433160 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data\") pod \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.433266 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle\") pod \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.433366 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data\") pod \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\" (UID: \"63c8f7bc-4162-4903-b3f9-96c8736a27b8\") " Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.444879 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "63c8f7bc-4162-4903-b3f9-96c8736a27b8" (UID: "63c8f7bc-4162-4903-b3f9-96c8736a27b8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.444943 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf" (OuterVolumeSpecName: "kube-api-access-vmmrf") pod "63c8f7bc-4162-4903-b3f9-96c8736a27b8" (UID: "63c8f7bc-4162-4903-b3f9-96c8736a27b8"). InnerVolumeSpecName "kube-api-access-vmmrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.460778 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63c8f7bc-4162-4903-b3f9-96c8736a27b8" (UID: "63c8f7bc-4162-4903-b3f9-96c8736a27b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.496454 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data" (OuterVolumeSpecName: "config-data") pod "63c8f7bc-4162-4903-b3f9-96c8736a27b8" (UID: "63c8f7bc-4162-4903-b3f9-96c8736a27b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.535516 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmmrf\" (UniqueName: \"kubernetes.io/projected/63c8f7bc-4162-4903-b3f9-96c8736a27b8-kube-api-access-vmmrf\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.535552 4733 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.535564 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.535577 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c8f7bc-4162-4903-b3f9-96c8736a27b8-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.954844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982"} Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.955966 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.957182 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ptbmt" event={"ID":"63c8f7bc-4162-4903-b3f9-96c8736a27b8","Type":"ContainerDied","Data":"0e53d30c9ba54d0bab85e4e6730952c64da04afa3c77dd75422e1f34b5188d78"} Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.957225 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e53d30c9ba54d0bab85e4e6730952c64da04afa3c77dd75422e1f34b5188d78" Mar 18 10:32:29 crc kubenswrapper[4733]: I0318 10:32:29.957272 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ptbmt" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.450717 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-z8m4g"] Mar 18 10:32:30 crc kubenswrapper[4733]: E0318 10:32:30.451011 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="dnsmasq-dns" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451022 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="dnsmasq-dns" Mar 18 10:32:30 crc kubenswrapper[4733]: E0318 10:32:30.451038 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="init" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451044 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="init" Mar 18 10:32:30 crc kubenswrapper[4733]: E0318 10:32:30.451057 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" containerName="glance-db-sync" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451063 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" containerName="glance-db-sync" Mar 18 10:32:30 crc kubenswrapper[4733]: E0318 10:32:30.451080 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" containerName="oc" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451085 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" containerName="oc" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451235 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bdf8dbb-ffe1-48d1-9c79-22e37dd882be" containerName="dnsmasq-dns" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451250 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" containerName="oc" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451260 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" containerName="glance-db-sync" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.451985 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.474700 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-z8m4g"] Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.650909 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.651215 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.651244 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spshf\" (UniqueName: \"kubernetes.io/projected/5fcd9264-61af-4872-82e6-8b0e1667ac70-kube-api-access-spshf\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.651371 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-config\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.651439 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.651475 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.752640 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.752690 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.752734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.752760 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.752777 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spshf\" (UniqueName: \"kubernetes.io/projected/5fcd9264-61af-4872-82e6-8b0e1667ac70-kube-api-access-spshf\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.753120 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-config\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.753622 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.753845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.753867 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.754786 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-config\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.755033 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fcd9264-61af-4872-82e6-8b0e1667ac70-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.781489 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spshf\" (UniqueName: \"kubernetes.io/projected/5fcd9264-61af-4872-82e6-8b0e1667ac70-kube-api-access-spshf\") pod \"dnsmasq-dns-5f59b8f679-z8m4g\" (UID: \"5fcd9264-61af-4872-82e6-8b0e1667ac70\") " pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.966418 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4" exitCode=0 Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.967304 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4"} Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.967348 4733 scope.go:117] "RemoveContainer" containerID="f53733af192d507492916cb8fdfb9e36a34a2d5b06777b3df798d1db42baebb1" Mar 18 10:32:30 crc kubenswrapper[4733]: I0318 10:32:30.967752 4733 scope.go:117] "RemoveContainer" containerID="c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4" Mar 18 10:32:30 crc kubenswrapper[4733]: E0318 10:32:30.967953 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:32:31 crc kubenswrapper[4733]: I0318 10:32:31.068330 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:31 crc kubenswrapper[4733]: I0318 10:32:31.512287 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-z8m4g"] Mar 18 10:32:31 crc kubenswrapper[4733]: W0318 10:32:31.514846 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fcd9264_61af_4872_82e6_8b0e1667ac70.slice/crio-24b45d20a646f19c4ac7c53f08aa1252d7641f341ab6088ea4d0e7ab19ced41f WatchSource:0}: Error finding container 24b45d20a646f19c4ac7c53f08aa1252d7641f341ab6088ea4d0e7ab19ced41f: Status 404 returned error can't find the container with id 24b45d20a646f19c4ac7c53f08aa1252d7641f341ab6088ea4d0e7ab19ced41f Mar 18 10:32:31 crc kubenswrapper[4733]: I0318 10:32:31.977484 4733 generic.go:334] "Generic (PLEG): container finished" podID="5fcd9264-61af-4872-82e6-8b0e1667ac70" containerID="683bc45d8d919be15f5e003415fd39aa3c57a2d498bf02cfe2428c0d955eea7e" exitCode=0 Mar 18 10:32:31 crc kubenswrapper[4733]: I0318 10:32:31.977548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" event={"ID":"5fcd9264-61af-4872-82e6-8b0e1667ac70","Type":"ContainerDied","Data":"683bc45d8d919be15f5e003415fd39aa3c57a2d498bf02cfe2428c0d955eea7e"} Mar 18 10:32:31 crc kubenswrapper[4733]: I0318 10:32:31.977907 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" event={"ID":"5fcd9264-61af-4872-82e6-8b0e1667ac70","Type":"ContainerStarted","Data":"24b45d20a646f19c4ac7c53f08aa1252d7641f341ab6088ea4d0e7ab19ced41f"} Mar 18 10:32:33 crc kubenswrapper[4733]: I0318 10:32:32.999716 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" event={"ID":"5fcd9264-61af-4872-82e6-8b0e1667ac70","Type":"ContainerStarted","Data":"4875c8e8342cbcaa2e5b0a8a9970108486fdbbadf86eefa8ef87d6df3340240a"} Mar 18 10:32:33 crc kubenswrapper[4733]: I0318 10:32:33.000376 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:33 crc kubenswrapper[4733]: I0318 10:32:33.029885 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" podStartSLOduration=3.02986627 podStartE2EDuration="3.02986627s" podCreationTimestamp="2026-03-18 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 10:32:33.021547405 +0000 UTC m=+1192.513281730" watchObservedRunningTime="2026-03-18 10:32:33.02986627 +0000 UTC m=+1192.521600585" Mar 18 10:32:34 crc kubenswrapper[4733]: I0318 10:32:34.009055 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982" exitCode=0 Mar 18 10:32:34 crc kubenswrapper[4733]: I0318 10:32:34.009138 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982"} Mar 18 10:32:34 crc kubenswrapper[4733]: I0318 10:32:34.009631 4733 scope.go:117] "RemoveContainer" containerID="9ba4505789b02a7aa27e40622e705a2188f60bebd64231f907e57dbee799f683" Mar 18 10:32:34 crc kubenswrapper[4733]: I0318 10:32:34.010403 4733 scope.go:117] "RemoveContainer" containerID="e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982" Mar 18 10:32:34 crc kubenswrapper[4733]: E0318 10:32:34.010666 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.071493 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-z8m4g" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.159475 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.159754 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="dnsmasq-dns" containerID="cri-o://42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db" gracePeriod=10 Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.685315 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772739 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbm6m\" (UniqueName: \"kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772774 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772824 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772857 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.772911 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb\") pod \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\" (UID: \"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d\") " Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.785334 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m" (OuterVolumeSpecName: "kube-api-access-zbm6m") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "kube-api-access-zbm6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.843014 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.843953 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.852970 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.857111 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.876394 4733 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.876459 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.876475 4733 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.876488 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbm6m\" (UniqueName: \"kubernetes.io/projected/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-kube-api-access-zbm6m\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.876502 4733 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.877571 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config" (OuterVolumeSpecName: "config") pod "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" (UID: "e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:32:41 crc kubenswrapper[4733]: I0318 10:32:41.977725 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d-config\") on node \"crc\" DevicePath \"\"" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.080984 4733 generic.go:334] "Generic (PLEG): container finished" podID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerID="42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db" exitCode=0 Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.081030 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.081031 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" event={"ID":"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d","Type":"ContainerDied","Data":"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db"} Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.081136 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-5vkkr" event={"ID":"e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d","Type":"ContainerDied","Data":"e1e841a260dd28bf7f12085c583e93ce7e0418239f8aa3a3290f316b2610e14f"} Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.081155 4733 scope.go:117] "RemoveContainer" containerID="42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.100159 4733 scope.go:117] "RemoveContainer" containerID="d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.114039 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.120813 4733 scope.go:117] "RemoveContainer" containerID="42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.121108 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-5vkkr"] Mar 18 10:32:42 crc kubenswrapper[4733]: E0318 10:32:42.121298 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db\": container with ID starting with 42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db not found: ID does not exist" containerID="42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.121337 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db"} err="failed to get container status \"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db\": rpc error: code = NotFound desc = could not find container \"42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db\": container with ID starting with 42e7ffa5d83fe25c846d3177b274430b818fc1f5e5b4e5a9ae3ffea34dec97db not found: ID does not exist" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.121368 4733 scope.go:117] "RemoveContainer" containerID="d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535" Mar 18 10:32:42 crc kubenswrapper[4733]: E0318 10:32:42.121664 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535\": container with ID starting with d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535 not found: ID does not exist" containerID="d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535" Mar 18 10:32:42 crc kubenswrapper[4733]: I0318 10:32:42.121693 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535"} err="failed to get container status \"d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535\": rpc error: code = NotFound desc = could not find container \"d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535\": container with ID starting with d08ced42dfa9a4c0f8b5c1fc1217494bf5cf7c8b883d4f35abfd833bac185535 not found: ID does not exist" Mar 18 10:32:43 crc kubenswrapper[4733]: I0318 10:32:43.176723 4733 scope.go:117] "RemoveContainer" containerID="c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4" Mar 18 10:32:43 crc kubenswrapper[4733]: E0318 10:32:43.177457 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:32:43 crc kubenswrapper[4733]: I0318 10:32:43.192980 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" path="/var/lib/kubelet/pods/e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d/volumes" Mar 18 10:32:43 crc kubenswrapper[4733]: I0318 10:32:43.571021 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:32:43 crc kubenswrapper[4733]: I0318 10:32:43.571119 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:32:48 crc kubenswrapper[4733]: I0318 10:32:48.175851 4733 scope.go:117] "RemoveContainer" containerID="e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982" Mar 18 10:32:48 crc kubenswrapper[4733]: E0318 10:32:48.176582 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:32:54 crc kubenswrapper[4733]: I0318 10:32:54.175669 4733 scope.go:117] "RemoveContainer" containerID="c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4" Mar 18 10:32:55 crc kubenswrapper[4733]: I0318 10:32:55.202537 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c"} Mar 18 10:32:55 crc kubenswrapper[4733]: I0318 10:32:55.203580 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:32:59 crc kubenswrapper[4733]: I0318 10:32:59.239632 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" exitCode=0 Mar 18 10:32:59 crc kubenswrapper[4733]: I0318 10:32:59.239764 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c"} Mar 18 10:32:59 crc kubenswrapper[4733]: I0318 10:32:59.240041 4733 scope.go:117] "RemoveContainer" containerID="c403cc91104f4f18606a75c2c7c0e5519b21ff1fde3dacb452abcf30617940a4" Mar 18 10:32:59 crc kubenswrapper[4733]: I0318 10:32:59.240765 4733 scope.go:117] "RemoveContainer" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" Mar 18 10:32:59 crc kubenswrapper[4733]: E0318 10:32:59.241111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:33:02 crc kubenswrapper[4733]: I0318 10:33:02.175984 4733 scope.go:117] "RemoveContainer" containerID="e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982" Mar 18 10:33:03 crc kubenswrapper[4733]: I0318 10:33:03.281881 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa"} Mar 18 10:33:03 crc kubenswrapper[4733]: I0318 10:33:03.282554 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:33:07 crc kubenswrapper[4733]: I0318 10:33:07.324852 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" exitCode=0 Mar 18 10:33:07 crc kubenswrapper[4733]: I0318 10:33:07.324911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa"} Mar 18 10:33:07 crc kubenswrapper[4733]: I0318 10:33:07.325331 4733 scope.go:117] "RemoveContainer" containerID="e897d31f6a9846ce7ce6f729eb4ad9ad29fd1f9d58f4b6a76aae048e641e8982" Mar 18 10:33:07 crc kubenswrapper[4733]: I0318 10:33:07.326727 4733 scope.go:117] "RemoveContainer" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" Mar 18 10:33:07 crc kubenswrapper[4733]: E0318 10:33:07.327166 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:33:08 crc kubenswrapper[4733]: I0318 10:33:08.570637 4733 scope.go:117] "RemoveContainer" containerID="73a17ce4bce512adc8ff4282e561fca0880aa24a1a28aaa52332d077a8673f8c" Mar 18 10:33:13 crc kubenswrapper[4733]: I0318 10:33:13.571884 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:33:13 crc kubenswrapper[4733]: I0318 10:33:13.572530 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:33:14 crc kubenswrapper[4733]: I0318 10:33:14.176445 4733 scope.go:117] "RemoveContainer" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" Mar 18 10:33:14 crc kubenswrapper[4733]: E0318 10:33:14.177107 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:33:22 crc kubenswrapper[4733]: I0318 10:33:22.175749 4733 scope.go:117] "RemoveContainer" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" Mar 18 10:33:22 crc kubenswrapper[4733]: E0318 10:33:22.176507 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:33:29 crc kubenswrapper[4733]: I0318 10:33:29.176618 4733 scope.go:117] "RemoveContainer" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" Mar 18 10:33:29 crc kubenswrapper[4733]: E0318 10:33:29.177531 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:33:36 crc kubenswrapper[4733]: I0318 10:33:36.175418 4733 scope.go:117] "RemoveContainer" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" Mar 18 10:33:36 crc kubenswrapper[4733]: E0318 10:33:36.176269 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:33:41 crc kubenswrapper[4733]: I0318 10:33:41.187832 4733 scope.go:117] "RemoveContainer" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" Mar 18 10:33:41 crc kubenswrapper[4733]: I0318 10:33:41.672691 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686"} Mar 18 10:33:41 crc kubenswrapper[4733]: I0318 10:33:41.674072 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:33:43 crc kubenswrapper[4733]: I0318 10:33:43.571355 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:33:43 crc kubenswrapper[4733]: I0318 10:33:43.571821 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:33:43 crc kubenswrapper[4733]: I0318 10:33:43.571906 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:33:43 crc kubenswrapper[4733]: I0318 10:33:43.573792 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:33:43 crc kubenswrapper[4733]: I0318 10:33:43.573916 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625" gracePeriod=600 Mar 18 10:33:44 crc kubenswrapper[4733]: I0318 10:33:44.710782 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625" exitCode=0 Mar 18 10:33:44 crc kubenswrapper[4733]: I0318 10:33:44.710901 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625"} Mar 18 10:33:44 crc kubenswrapper[4733]: I0318 10:33:44.711176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9"} Mar 18 10:33:44 crc kubenswrapper[4733]: I0318 10:33:44.711216 4733 scope.go:117] "RemoveContainer" containerID="2a78644e078fbb319d0fc66d47cfb2501076e4fd678ad793e791ddb4f3d3ee96" Mar 18 10:33:45 crc kubenswrapper[4733]: I0318 10:33:45.733456 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" exitCode=0 Mar 18 10:33:45 crc kubenswrapper[4733]: I0318 10:33:45.733539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686"} Mar 18 10:33:45 crc kubenswrapper[4733]: I0318 10:33:45.734029 4733 scope.go:117] "RemoveContainer" containerID="bdf571c67b493e3fd2b9642e6dd66519b598c153da1f077075fd25aebffa1e9c" Mar 18 10:33:45 crc kubenswrapper[4733]: I0318 10:33:45.735481 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:33:45 crc kubenswrapper[4733]: E0318 10:33:45.736178 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:33:49 crc kubenswrapper[4733]: I0318 10:33:49.175960 4733 scope.go:117] "RemoveContainer" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" Mar 18 10:33:49 crc kubenswrapper[4733]: I0318 10:33:49.784683 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00"} Mar 18 10:33:49 crc kubenswrapper[4733]: I0318 10:33:49.785578 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:33:53 crc kubenswrapper[4733]: I0318 10:33:53.822208 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" exitCode=0 Mar 18 10:33:53 crc kubenswrapper[4733]: I0318 10:33:53.822322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00"} Mar 18 10:33:53 crc kubenswrapper[4733]: I0318 10:33:53.822834 4733 scope.go:117] "RemoveContainer" containerID="404382805ba91938d5973ffc7857ed67b92775fae6cf128d8db979d4adcb6eaa" Mar 18 10:33:53 crc kubenswrapper[4733]: I0318 10:33:53.823680 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:33:53 crc kubenswrapper[4733]: E0318 10:33:53.824007 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:33:58 crc kubenswrapper[4733]: I0318 10:33:58.176402 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:33:58 crc kubenswrapper[4733]: E0318 10:33:58.177111 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.157638 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563834-7bqxq"] Mar 18 10:34:00 crc kubenswrapper[4733]: E0318 10:34:00.158538 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="dnsmasq-dns" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.158559 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="dnsmasq-dns" Mar 18 10:34:00 crc kubenswrapper[4733]: E0318 10:34:00.158580 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="init" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.158592 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="init" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.158898 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c771ba-dbb2-470b-b19c-8c8fdefbdd6d" containerName="dnsmasq-dns" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.159807 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.165036 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.167710 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.168829 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.170057 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563834-7bqxq"] Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.201707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crlqs\" (UniqueName: \"kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs\") pod \"auto-csr-approver-29563834-7bqxq\" (UID: \"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316\") " pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.304498 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crlqs\" (UniqueName: \"kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs\") pod \"auto-csr-approver-29563834-7bqxq\" (UID: \"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316\") " pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.344269 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crlqs\" (UniqueName: \"kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs\") pod \"auto-csr-approver-29563834-7bqxq\" (UID: \"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316\") " pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:00 crc kubenswrapper[4733]: I0318 10:34:00.513896 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:01 crc kubenswrapper[4733]: I0318 10:34:01.632276 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563834-7bqxq"] Mar 18 10:34:01 crc kubenswrapper[4733]: W0318 10:34:01.635411 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc0f56eb_6c6a_49bf_9a12_ef5f2dd95316.slice/crio-29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17 WatchSource:0}: Error finding container 29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17: Status 404 returned error can't find the container with id 29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17 Mar 18 10:34:01 crc kubenswrapper[4733]: I0318 10:34:01.905231 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" event={"ID":"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316","Type":"ContainerStarted","Data":"29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17"} Mar 18 10:34:03 crc kubenswrapper[4733]: I0318 10:34:03.930377 4733 generic.go:334] "Generic (PLEG): container finished" podID="dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" containerID="2746f736c334d9ac3079e5dc9b5db5929c610a6933e47547e58536ec78e443c9" exitCode=0 Mar 18 10:34:03 crc kubenswrapper[4733]: I0318 10:34:03.930472 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" event={"ID":"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316","Type":"ContainerDied","Data":"2746f736c334d9ac3079e5dc9b5db5929c610a6933e47547e58536ec78e443c9"} Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.350340 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.389032 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crlqs\" (UniqueName: \"kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs\") pod \"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316\" (UID: \"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316\") " Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.399471 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs" (OuterVolumeSpecName: "kube-api-access-crlqs") pod "dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" (UID: "dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316"). InnerVolumeSpecName "kube-api-access-crlqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.490305 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crlqs\" (UniqueName: \"kubernetes.io/projected/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316-kube-api-access-crlqs\") on node \"crc\" DevicePath \"\"" Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.965677 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" event={"ID":"dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316","Type":"ContainerDied","Data":"29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17"} Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.966165 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b17b8ff0f2a892cfd860135ff82140434dd4ba914f10aa584db885c8ec2e17" Mar 18 10:34:05 crc kubenswrapper[4733]: I0318 10:34:05.966041 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563834-7bqxq" Mar 18 10:34:06 crc kubenswrapper[4733]: I0318 10:34:06.434054 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563828-zczv7"] Mar 18 10:34:06 crc kubenswrapper[4733]: I0318 10:34:06.448233 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563828-zczv7"] Mar 18 10:34:07 crc kubenswrapper[4733]: I0318 10:34:07.175138 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:34:07 crc kubenswrapper[4733]: E0318 10:34:07.175394 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:34:07 crc kubenswrapper[4733]: I0318 10:34:07.188999 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68574d72-725d-48c2-b645-bd83dcccbf80" path="/var/lib/kubelet/pods/68574d72-725d-48c2-b645-bd83dcccbf80/volumes" Mar 18 10:34:08 crc kubenswrapper[4733]: I0318 10:34:08.687280 4733 scope.go:117] "RemoveContainer" containerID="f4a3549ea82cce03bd994263d641938a407bdfdc2f86792bccee0b653493614d" Mar 18 10:34:12 crc kubenswrapper[4733]: I0318 10:34:12.177602 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:34:12 crc kubenswrapper[4733]: E0318 10:34:12.178650 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:34:20 crc kubenswrapper[4733]: I0318 10:34:20.176083 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:34:20 crc kubenswrapper[4733]: E0318 10:34:20.176929 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:34:23 crc kubenswrapper[4733]: I0318 10:34:23.176370 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:34:23 crc kubenswrapper[4733]: E0318 10:34:23.177308 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:34:35 crc kubenswrapper[4733]: I0318 10:34:35.177413 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:34:35 crc kubenswrapper[4733]: E0318 10:34:35.179051 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:34:36 crc kubenswrapper[4733]: I0318 10:34:36.175560 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:34:36 crc kubenswrapper[4733]: E0318 10:34:36.176063 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:34:47 crc kubenswrapper[4733]: I0318 10:34:47.175606 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:34:47 crc kubenswrapper[4733]: E0318 10:34:47.176871 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:34:50 crc kubenswrapper[4733]: I0318 10:34:50.175584 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:34:50 crc kubenswrapper[4733]: E0318 10:34:50.176171 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:00 crc kubenswrapper[4733]: I0318 10:35:00.175976 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:35:00 crc kubenswrapper[4733]: E0318 10:35:00.177096 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:35:04 crc kubenswrapper[4733]: I0318 10:35:04.194146 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:35:04 crc kubenswrapper[4733]: E0318 10:35:04.195258 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:11 crc kubenswrapper[4733]: I0318 10:35:11.184555 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:35:11 crc kubenswrapper[4733]: I0318 10:35:11.655012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a"} Mar 18 10:35:11 crc kubenswrapper[4733]: I0318 10:35:11.655681 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.176510 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.696863 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e"} Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.697183 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.701180 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" exitCode=0 Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.701323 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a"} Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.701442 4733 scope.go:117] "RemoveContainer" containerID="693e6eba0eed87d5064eb695aef9f113ca2e6fa1b8fe4241bf5171215cf4e686" Mar 18 10:35:15 crc kubenswrapper[4733]: I0318 10:35:15.702137 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:35:15 crc kubenswrapper[4733]: E0318 10:35:15.702568 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:35:19 crc kubenswrapper[4733]: I0318 10:35:19.751477 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" exitCode=0 Mar 18 10:35:19 crc kubenswrapper[4733]: I0318 10:35:19.751631 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e"} Mar 18 10:35:19 crc kubenswrapper[4733]: I0318 10:35:19.752348 4733 scope.go:117] "RemoveContainer" containerID="f53819fc80d135628a27c9199c900935f63ae50e8bb03f43fb957f7ef27dfd00" Mar 18 10:35:19 crc kubenswrapper[4733]: I0318 10:35:19.753641 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:35:19 crc kubenswrapper[4733]: E0318 10:35:19.754023 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:30 crc kubenswrapper[4733]: I0318 10:35:30.175796 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:35:30 crc kubenswrapper[4733]: I0318 10:35:30.178339 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:35:30 crc kubenswrapper[4733]: E0318 10:35:30.178826 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:30 crc kubenswrapper[4733]: E0318 10:35:30.178845 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:35:41 crc kubenswrapper[4733]: I0318 10:35:41.182249 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:35:41 crc kubenswrapper[4733]: E0318 10:35:41.183093 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:43 crc kubenswrapper[4733]: I0318 10:35:43.571044 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:35:43 crc kubenswrapper[4733]: I0318 10:35:43.571527 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:35:44 crc kubenswrapper[4733]: I0318 10:35:44.175605 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:35:44 crc kubenswrapper[4733]: E0318 10:35:44.176506 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:35:52 crc kubenswrapper[4733]: I0318 10:35:52.175597 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:35:52 crc kubenswrapper[4733]: E0318 10:35:52.176639 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:35:58 crc kubenswrapper[4733]: I0318 10:35:58.175554 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:35:58 crc kubenswrapper[4733]: E0318 10:35:58.176344 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.149229 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563836-4x58h"] Mar 18 10:36:00 crc kubenswrapper[4733]: E0318 10:36:00.149954 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" containerName="oc" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.149971 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" containerName="oc" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.150218 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" containerName="oc" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.150795 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.153355 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.153407 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.154667 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.156978 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563836-4x58h"] Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.263636 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c46cx\" (UniqueName: \"kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx\") pod \"auto-csr-approver-29563836-4x58h\" (UID: \"d31d006c-81a6-4bbb-a44a-fda966944372\") " pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.366031 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c46cx\" (UniqueName: \"kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx\") pod \"auto-csr-approver-29563836-4x58h\" (UID: \"d31d006c-81a6-4bbb-a44a-fda966944372\") " pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.386619 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c46cx\" (UniqueName: \"kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx\") pod \"auto-csr-approver-29563836-4x58h\" (UID: \"d31d006c-81a6-4bbb-a44a-fda966944372\") " pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.472967 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:00 crc kubenswrapper[4733]: I0318 10:36:00.940454 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563836-4x58h"] Mar 18 10:36:01 crc kubenswrapper[4733]: I0318 10:36:01.554127 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563836-4x58h" event={"ID":"d31d006c-81a6-4bbb-a44a-fda966944372","Type":"ContainerStarted","Data":"450d93d363176e50bafab37a68c6e96830ec396aca9eb8497b1b034d1c6120d3"} Mar 18 10:36:02 crc kubenswrapper[4733]: I0318 10:36:02.564875 4733 generic.go:334] "Generic (PLEG): container finished" podID="d31d006c-81a6-4bbb-a44a-fda966944372" containerID="4d1446897edd6664fd044255842a94c2933d1bab0fe0d09f54123f9a53833063" exitCode=0 Mar 18 10:36:02 crc kubenswrapper[4733]: I0318 10:36:02.564974 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563836-4x58h" event={"ID":"d31d006c-81a6-4bbb-a44a-fda966944372","Type":"ContainerDied","Data":"4d1446897edd6664fd044255842a94c2933d1bab0fe0d09f54123f9a53833063"} Mar 18 10:36:03 crc kubenswrapper[4733]: I0318 10:36:03.177304 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:36:03 crc kubenswrapper[4733]: E0318 10:36:03.177626 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:36:03 crc kubenswrapper[4733]: I0318 10:36:03.911078 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.029757 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c46cx\" (UniqueName: \"kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx\") pod \"d31d006c-81a6-4bbb-a44a-fda966944372\" (UID: \"d31d006c-81a6-4bbb-a44a-fda966944372\") " Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.037011 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx" (OuterVolumeSpecName: "kube-api-access-c46cx") pod "d31d006c-81a6-4bbb-a44a-fda966944372" (UID: "d31d006c-81a6-4bbb-a44a-fda966944372"). InnerVolumeSpecName "kube-api-access-c46cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.131273 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c46cx\" (UniqueName: \"kubernetes.io/projected/d31d006c-81a6-4bbb-a44a-fda966944372-kube-api-access-c46cx\") on node \"crc\" DevicePath \"\"" Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.581954 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563836-4x58h" event={"ID":"d31d006c-81a6-4bbb-a44a-fda966944372","Type":"ContainerDied","Data":"450d93d363176e50bafab37a68c6e96830ec396aca9eb8497b1b034d1c6120d3"} Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.581998 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="450d93d363176e50bafab37a68c6e96830ec396aca9eb8497b1b034d1c6120d3" Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.582050 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563836-4x58h" Mar 18 10:36:04 crc kubenswrapper[4733]: I0318 10:36:04.991521 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563830-2qqd2"] Mar 18 10:36:05 crc kubenswrapper[4733]: I0318 10:36:05.002392 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563830-2qqd2"] Mar 18 10:36:05 crc kubenswrapper[4733]: I0318 10:36:05.200295 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c8eb139-c576-4ab3-8c2b-a309f3aa4a35" path="/var/lib/kubelet/pods/3c8eb139-c576-4ab3-8c2b-a309f3aa4a35/volumes" Mar 18 10:36:08 crc kubenswrapper[4733]: I0318 10:36:08.807150 4733 scope.go:117] "RemoveContainer" containerID="be977da7d932bb787db9cafb1727d3bd50b5e03495d1e8a82c232ed7c66e241e" Mar 18 10:36:13 crc kubenswrapper[4733]: I0318 10:36:13.176735 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:36:13 crc kubenswrapper[4733]: E0318 10:36:13.177785 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:36:13 crc kubenswrapper[4733]: I0318 10:36:13.571722 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:36:13 crc kubenswrapper[4733]: I0318 10:36:13.571831 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:36:14 crc kubenswrapper[4733]: I0318 10:36:14.176059 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:36:14 crc kubenswrapper[4733]: E0318 10:36:14.176893 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:36:25 crc kubenswrapper[4733]: I0318 10:36:25.176921 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:36:25 crc kubenswrapper[4733]: E0318 10:36:25.189522 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:36:27 crc kubenswrapper[4733]: I0318 10:36:27.175941 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:36:27 crc kubenswrapper[4733]: E0318 10:36:27.176636 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:36:37 crc kubenswrapper[4733]: I0318 10:36:37.176043 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:36:37 crc kubenswrapper[4733]: E0318 10:36:37.178739 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:36:41 crc kubenswrapper[4733]: I0318 10:36:41.176297 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:36:41 crc kubenswrapper[4733]: E0318 10:36:41.177743 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:36:43 crc kubenswrapper[4733]: I0318 10:36:43.571855 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:36:43 crc kubenswrapper[4733]: I0318 10:36:43.572337 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:36:43 crc kubenswrapper[4733]: I0318 10:36:43.572412 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:36:43 crc kubenswrapper[4733]: I0318 10:36:43.573400 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:36:43 crc kubenswrapper[4733]: I0318 10:36:43.573500 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9" gracePeriod=600 Mar 18 10:36:44 crc kubenswrapper[4733]: I0318 10:36:44.021808 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9" exitCode=0 Mar 18 10:36:44 crc kubenswrapper[4733]: I0318 10:36:44.022025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9"} Mar 18 10:36:44 crc kubenswrapper[4733]: I0318 10:36:44.022219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637"} Mar 18 10:36:44 crc kubenswrapper[4733]: I0318 10:36:44.022276 4733 scope.go:117] "RemoveContainer" containerID="345f1c51e0b2f38e27fd31ce4a7323d51ffa4b8419f456177dd8653558afb625" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.085860 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:36:50 crc kubenswrapper[4733]: E0318 10:36:50.088068 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31d006c-81a6-4bbb-a44a-fda966944372" containerName="oc" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.088101 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31d006c-81a6-4bbb-a44a-fda966944372" containerName="oc" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.089037 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31d006c-81a6-4bbb-a44a-fda966944372" containerName="oc" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.102608 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.123082 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.202473 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.202886 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.202996 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46xhq\" (UniqueName: \"kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.305009 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46xhq\" (UniqueName: \"kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.307160 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.307845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.308665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.309363 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.332504 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46xhq\" (UniqueName: \"kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq\") pod \"community-operators-g9kvf\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.447943 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:36:50 crc kubenswrapper[4733]: I0318 10:36:50.960714 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:36:51 crc kubenswrapper[4733]: I0318 10:36:51.125675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerStarted","Data":"77b5823cb17c9bd5ad620ae110d159272c1807518ae06bbaa2150540125e3326"} Mar 18 10:36:51 crc kubenswrapper[4733]: I0318 10:36:51.189157 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:36:51 crc kubenswrapper[4733]: E0318 10:36:51.189830 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:36:52 crc kubenswrapper[4733]: I0318 10:36:52.138710 4733 generic.go:334] "Generic (PLEG): container finished" podID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerID="c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867" exitCode=0 Mar 18 10:36:52 crc kubenswrapper[4733]: I0318 10:36:52.138785 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerDied","Data":"c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867"} Mar 18 10:36:52 crc kubenswrapper[4733]: I0318 10:36:52.177005 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:36:52 crc kubenswrapper[4733]: E0318 10:36:52.177430 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.152227 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerStarted","Data":"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37"} Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.674853 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.682956 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.688427 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.773133 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mspkt\" (UniqueName: \"kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.773409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.773606 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.874883 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mspkt\" (UniqueName: \"kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.875077 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.875779 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.875939 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.876436 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:53 crc kubenswrapper[4733]: I0318 10:36:53.902112 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mspkt\" (UniqueName: \"kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt\") pod \"redhat-marketplace-mzcst\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:54 crc kubenswrapper[4733]: I0318 10:36:54.010764 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:36:54 crc kubenswrapper[4733]: I0318 10:36:54.178490 4733 generic.go:334] "Generic (PLEG): container finished" podID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerID="0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37" exitCode=0 Mar 18 10:36:54 crc kubenswrapper[4733]: I0318 10:36:54.178548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerDied","Data":"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37"} Mar 18 10:36:54 crc kubenswrapper[4733]: I0318 10:36:54.509670 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:36:55 crc kubenswrapper[4733]: I0318 10:36:55.197933 4733 generic.go:334] "Generic (PLEG): container finished" podID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerID="43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b" exitCode=0 Mar 18 10:36:55 crc kubenswrapper[4733]: I0318 10:36:55.198038 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerDied","Data":"43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b"} Mar 18 10:36:55 crc kubenswrapper[4733]: I0318 10:36:55.198377 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerStarted","Data":"733fa8b795ed27071ff92b08dafee2478465f1de0391a9876c4ef3a903350321"} Mar 18 10:36:55 crc kubenswrapper[4733]: I0318 10:36:55.201463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerStarted","Data":"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0"} Mar 18 10:36:55 crc kubenswrapper[4733]: I0318 10:36:55.245046 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g9kvf" podStartSLOduration=2.731001655 podStartE2EDuration="5.245021351s" podCreationTimestamp="2026-03-18 10:36:50 +0000 UTC" firstStartedPulling="2026-03-18 10:36:52.140860408 +0000 UTC m=+1451.632594763" lastFinishedPulling="2026-03-18 10:36:54.654880134 +0000 UTC m=+1454.146614459" observedRunningTime="2026-03-18 10:36:55.235597225 +0000 UTC m=+1454.727331550" watchObservedRunningTime="2026-03-18 10:36:55.245021351 +0000 UTC m=+1454.736755696" Mar 18 10:36:56 crc kubenswrapper[4733]: I0318 10:36:56.210930 4733 generic.go:334] "Generic (PLEG): container finished" podID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerID="1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e" exitCode=0 Mar 18 10:36:56 crc kubenswrapper[4733]: I0318 10:36:56.211029 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerDied","Data":"1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e"} Mar 18 10:36:57 crc kubenswrapper[4733]: I0318 10:36:57.236586 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerStarted","Data":"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7"} Mar 18 10:36:57 crc kubenswrapper[4733]: I0318 10:36:57.269010 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mzcst" podStartSLOduration=2.884084295 podStartE2EDuration="4.26899436s" podCreationTimestamp="2026-03-18 10:36:53 +0000 UTC" firstStartedPulling="2026-03-18 10:36:55.201060743 +0000 UTC m=+1454.692795068" lastFinishedPulling="2026-03-18 10:36:56.585970778 +0000 UTC m=+1456.077705133" observedRunningTime="2026-03-18 10:36:57.263614658 +0000 UTC m=+1456.755348983" watchObservedRunningTime="2026-03-18 10:36:57.26899436 +0000 UTC m=+1456.760728675" Mar 18 10:37:00 crc kubenswrapper[4733]: I0318 10:37:00.449574 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:00 crc kubenswrapper[4733]: I0318 10:37:00.450868 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:00 crc kubenswrapper[4733]: I0318 10:37:00.527543 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:01 crc kubenswrapper[4733]: I0318 10:37:01.310602 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:02 crc kubenswrapper[4733]: I0318 10:37:02.859980 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:37:03 crc kubenswrapper[4733]: I0318 10:37:03.176095 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:37:03 crc kubenswrapper[4733]: E0318 10:37:03.176629 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.011527 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.011605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.069000 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.295847 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g9kvf" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="registry-server" containerID="cri-o://137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0" gracePeriod=2 Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.371728 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.810769 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.960811 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities\") pod \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.961010 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46xhq\" (UniqueName: \"kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq\") pod \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.961036 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content\") pod \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\" (UID: \"a2b2cea8-3250-49e7-9afb-8cd8adfbc175\") " Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.962083 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities" (OuterVolumeSpecName: "utilities") pod "a2b2cea8-3250-49e7-9afb-8cd8adfbc175" (UID: "a2b2cea8-3250-49e7-9afb-8cd8adfbc175"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:37:04 crc kubenswrapper[4733]: I0318 10:37:04.971152 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq" (OuterVolumeSpecName: "kube-api-access-46xhq") pod "a2b2cea8-3250-49e7-9afb-8cd8adfbc175" (UID: "a2b2cea8-3250-49e7-9afb-8cd8adfbc175"). InnerVolumeSpecName "kube-api-access-46xhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.059434 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2b2cea8-3250-49e7-9afb-8cd8adfbc175" (UID: "a2b2cea8-3250-49e7-9afb-8cd8adfbc175"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.063635 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.063699 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46xhq\" (UniqueName: \"kubernetes.io/projected/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-kube-api-access-46xhq\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.063721 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2b2cea8-3250-49e7-9afb-8cd8adfbc175-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.175589 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:37:05 crc kubenswrapper[4733]: E0318 10:37:05.176081 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.312096 4733 generic.go:334] "Generic (PLEG): container finished" podID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerID="137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0" exitCode=0 Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.312242 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9kvf" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.312236 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerDied","Data":"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0"} Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.312406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9kvf" event={"ID":"a2b2cea8-3250-49e7-9afb-8cd8adfbc175","Type":"ContainerDied","Data":"77b5823cb17c9bd5ad620ae110d159272c1807518ae06bbaa2150540125e3326"} Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.312451 4733 scope.go:117] "RemoveContainer" containerID="137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.356464 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.362089 4733 scope.go:117] "RemoveContainer" containerID="0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.363845 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g9kvf"] Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.393149 4733 scope.go:117] "RemoveContainer" containerID="c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.434562 4733 scope.go:117] "RemoveContainer" containerID="137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0" Mar 18 10:37:05 crc kubenswrapper[4733]: E0318 10:37:05.435323 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0\": container with ID starting with 137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0 not found: ID does not exist" containerID="137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.435366 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0"} err="failed to get container status \"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0\": rpc error: code = NotFound desc = could not find container \"137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0\": container with ID starting with 137f4b5cd6409671241846735f9faafc7b02148bf4e909270882de413e75c2d0 not found: ID does not exist" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.435392 4733 scope.go:117] "RemoveContainer" containerID="0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37" Mar 18 10:37:05 crc kubenswrapper[4733]: E0318 10:37:05.435935 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37\": container with ID starting with 0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37 not found: ID does not exist" containerID="0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.436005 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37"} err="failed to get container status \"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37\": rpc error: code = NotFound desc = could not find container \"0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37\": container with ID starting with 0182ab6d575fb1256d6047471250a820525e208bd74c344e9e4505523d32cf37 not found: ID does not exist" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.436053 4733 scope.go:117] "RemoveContainer" containerID="c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867" Mar 18 10:37:05 crc kubenswrapper[4733]: E0318 10:37:05.438447 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867\": container with ID starting with c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867 not found: ID does not exist" containerID="c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867" Mar 18 10:37:05 crc kubenswrapper[4733]: I0318 10:37:05.438527 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867"} err="failed to get container status \"c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867\": rpc error: code = NotFound desc = could not find container \"c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867\": container with ID starting with c5a21fa0cbceaea43ca29ac1a7ba14c49e0499e62c04076b20c4a0dea541c867 not found: ID does not exist" Mar 18 10:37:06 crc kubenswrapper[4733]: I0318 10:37:06.460661 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:37:06 crc kubenswrapper[4733]: I0318 10:37:06.461312 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mzcst" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="registry-server" containerID="cri-o://e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7" gracePeriod=2 Mar 18 10:37:06 crc kubenswrapper[4733]: I0318 10:37:06.973940 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.102092 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mspkt\" (UniqueName: \"kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt\") pod \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.102171 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities\") pod \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.102365 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content\") pod \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\" (UID: \"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9\") " Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.103958 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities" (OuterVolumeSpecName: "utilities") pod "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" (UID: "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.112675 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt" (OuterVolumeSpecName: "kube-api-access-mspkt") pod "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" (UID: "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9"). InnerVolumeSpecName "kube-api-access-mspkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.155265 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" (UID: "6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.190704 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" path="/var/lib/kubelet/pods/a2b2cea8-3250-49e7-9afb-8cd8adfbc175/volumes" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.204595 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.204673 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mspkt\" (UniqueName: \"kubernetes.io/projected/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-kube-api-access-mspkt\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.204702 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.344083 4733 generic.go:334] "Generic (PLEG): container finished" podID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerID="e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7" exitCode=0 Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.344143 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerDied","Data":"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7"} Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.344182 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzcst" event={"ID":"6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9","Type":"ContainerDied","Data":"733fa8b795ed27071ff92b08dafee2478465f1de0391a9876c4ef3a903350321"} Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.344242 4733 scope.go:117] "RemoveContainer" containerID="e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.344430 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzcst" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.385502 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.387054 4733 scope.go:117] "RemoveContainer" containerID="1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.394136 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzcst"] Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.416554 4733 scope.go:117] "RemoveContainer" containerID="43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.457496 4733 scope.go:117] "RemoveContainer" containerID="e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7" Mar 18 10:37:07 crc kubenswrapper[4733]: E0318 10:37:07.458025 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7\": container with ID starting with e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7 not found: ID does not exist" containerID="e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.458085 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7"} err="failed to get container status \"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7\": rpc error: code = NotFound desc = could not find container \"e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7\": container with ID starting with e01b9cb5a9a576b6af51a3c336323b76b216d79f06485e7f64ace3d275aedff7 not found: ID does not exist" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.458117 4733 scope.go:117] "RemoveContainer" containerID="1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e" Mar 18 10:37:07 crc kubenswrapper[4733]: E0318 10:37:07.458655 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e\": container with ID starting with 1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e not found: ID does not exist" containerID="1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.458711 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e"} err="failed to get container status \"1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e\": rpc error: code = NotFound desc = could not find container \"1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e\": container with ID starting with 1d3c68b39fae2dbff6428ba8437ca1b3885a84632766d99dab081c76cfafac7e not found: ID does not exist" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.458750 4733 scope.go:117] "RemoveContainer" containerID="43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b" Mar 18 10:37:07 crc kubenswrapper[4733]: E0318 10:37:07.459299 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b\": container with ID starting with 43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b not found: ID does not exist" containerID="43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b" Mar 18 10:37:07 crc kubenswrapper[4733]: I0318 10:37:07.459359 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b"} err="failed to get container status \"43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b\": rpc error: code = NotFound desc = could not find container \"43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b\": container with ID starting with 43bc7fdfbeffa77f19586c47b1204bb99968bbe1c8c66ec263c804c270f6200b not found: ID does not exist" Mar 18 10:37:09 crc kubenswrapper[4733]: I0318 10:37:09.192184 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" path="/var/lib/kubelet/pods/6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9/volumes" Mar 18 10:37:17 crc kubenswrapper[4733]: I0318 10:37:17.174973 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:37:17 crc kubenswrapper[4733]: E0318 10:37:17.175705 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:37:17 crc kubenswrapper[4733]: I0318 10:37:17.175850 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:37:17 crc kubenswrapper[4733]: E0318 10:37:17.176128 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:37:28 crc kubenswrapper[4733]: I0318 10:37:28.175859 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:37:28 crc kubenswrapper[4733]: E0318 10:37:28.177157 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:37:29 crc kubenswrapper[4733]: I0318 10:37:29.176436 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:37:29 crc kubenswrapper[4733]: E0318 10:37:29.176915 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:37:44 crc kubenswrapper[4733]: I0318 10:37:44.175663 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:37:44 crc kubenswrapper[4733]: I0318 10:37:44.176343 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:37:44 crc kubenswrapper[4733]: E0318 10:37:44.176687 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:37:44 crc kubenswrapper[4733]: E0318 10:37:44.176759 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.566180 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567609 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="extract-utilities" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567647 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="extract-utilities" Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567693 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="extract-content" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567712 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="extract-content" Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567766 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="extract-content" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567785 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="extract-content" Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567818 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567836 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567861 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567877 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: E0318 10:37:54.567905 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="extract-utilities" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.567923 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="extract-utilities" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.570177 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a9e3c2c-c4de-4719-88bb-6392a6f7f0e9" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.570308 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b2cea8-3250-49e7-9afb-8cd8adfbc175" containerName="registry-server" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.572753 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.585216 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.730709 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.731213 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjpzr\" (UniqueName: \"kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.731302 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.833933 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjpzr\" (UniqueName: \"kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.834000 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.834129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.835045 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.835151 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.861123 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjpzr\" (UniqueName: \"kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr\") pod \"redhat-operators-wxvvt\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:54 crc kubenswrapper[4733]: I0318 10:37:54.949440 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:37:55 crc kubenswrapper[4733]: I0318 10:37:55.398140 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:37:55 crc kubenswrapper[4733]: I0318 10:37:55.848732 4733 generic.go:334] "Generic (PLEG): container finished" podID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerID="7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e" exitCode=0 Mar 18 10:37:55 crc kubenswrapper[4733]: I0318 10:37:55.848801 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerDied","Data":"7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e"} Mar 18 10:37:55 crc kubenswrapper[4733]: I0318 10:37:55.848840 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerStarted","Data":"98f9eb05565700e8349c0932c41e77ad80e45cc728fe839b9e5a63014e0f9e85"} Mar 18 10:37:55 crc kubenswrapper[4733]: I0318 10:37:55.852149 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:37:57 crc kubenswrapper[4733]: I0318 10:37:57.176419 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:37:57 crc kubenswrapper[4733]: E0318 10:37:57.177795 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 2m40s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:37:57 crc kubenswrapper[4733]: I0318 10:37:57.882227 4733 generic.go:334] "Generic (PLEG): container finished" podID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerID="c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74" exitCode=0 Mar 18 10:37:57 crc kubenswrapper[4733]: I0318 10:37:57.882268 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerDied","Data":"c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74"} Mar 18 10:37:58 crc kubenswrapper[4733]: I0318 10:37:58.175986 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:37:58 crc kubenswrapper[4733]: I0318 10:37:58.892058 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a"} Mar 18 10:37:58 crc kubenswrapper[4733]: I0318 10:37:58.893146 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:37:58 crc kubenswrapper[4733]: I0318 10:37:58.894517 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerStarted","Data":"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa"} Mar 18 10:37:58 crc kubenswrapper[4733]: I0318 10:37:58.933466 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wxvvt" podStartSLOduration=2.286969509 podStartE2EDuration="4.933435025s" podCreationTimestamp="2026-03-18 10:37:54 +0000 UTC" firstStartedPulling="2026-03-18 10:37:55.85194359 +0000 UTC m=+1515.343677915" lastFinishedPulling="2026-03-18 10:37:58.498409076 +0000 UTC m=+1517.990143431" observedRunningTime="2026-03-18 10:37:58.930374029 +0000 UTC m=+1518.422108354" watchObservedRunningTime="2026-03-18 10:37:58.933435025 +0000 UTC m=+1518.425169350" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.150338 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563838-6fbsd"] Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.151528 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.154470 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.154602 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.169256 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563838-6fbsd"] Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.170134 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.231387 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcj82\" (UniqueName: \"kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82\") pod \"auto-csr-approver-29563838-6fbsd\" (UID: \"3959ab36-a688-40ac-b70b-b3cc35b1d7a1\") " pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.333111 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcj82\" (UniqueName: \"kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82\") pod \"auto-csr-approver-29563838-6fbsd\" (UID: \"3959ab36-a688-40ac-b70b-b3cc35b1d7a1\") " pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.371527 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcj82\" (UniqueName: \"kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82\") pod \"auto-csr-approver-29563838-6fbsd\" (UID: \"3959ab36-a688-40ac-b70b-b3cc35b1d7a1\") " pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:00 crc kubenswrapper[4733]: I0318 10:38:00.486446 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:00 crc kubenswrapper[4733]: W0318 10:38:00.993280 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3959ab36_a688_40ac_b70b_b3cc35b1d7a1.slice/crio-767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d WatchSource:0}: Error finding container 767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d: Status 404 returned error can't find the container with id 767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d Mar 18 10:38:01 crc kubenswrapper[4733]: I0318 10:38:01.006609 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563838-6fbsd"] Mar 18 10:38:01 crc kubenswrapper[4733]: I0318 10:38:01.922971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" event={"ID":"3959ab36-a688-40ac-b70b-b3cc35b1d7a1","Type":"ContainerStarted","Data":"767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d"} Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.941285 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" exitCode=0 Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.941496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a"} Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.941874 4733 scope.go:117] "RemoveContainer" containerID="245acab52e36967117888a03ca9615fc134fa986b5326378b27571bcb153bf6a" Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.942954 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:38:02 crc kubenswrapper[4733]: E0318 10:38:02.943455 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.947710 4733 generic.go:334] "Generic (PLEG): container finished" podID="3959ab36-a688-40ac-b70b-b3cc35b1d7a1" containerID="ac18342a3539a4f4eb0b18430ab3c33bd2af4e21dfc3695dc34573a145ad949d" exitCode=0 Mar 18 10:38:02 crc kubenswrapper[4733]: I0318 10:38:02.947783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" event={"ID":"3959ab36-a688-40ac-b70b-b3cc35b1d7a1","Type":"ContainerDied","Data":"ac18342a3539a4f4eb0b18430ab3c33bd2af4e21dfc3695dc34573a145ad949d"} Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.299127 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.423334 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcj82\" (UniqueName: \"kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82\") pod \"3959ab36-a688-40ac-b70b-b3cc35b1d7a1\" (UID: \"3959ab36-a688-40ac-b70b-b3cc35b1d7a1\") " Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.433750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82" (OuterVolumeSpecName: "kube-api-access-tcj82") pod "3959ab36-a688-40ac-b70b-b3cc35b1d7a1" (UID: "3959ab36-a688-40ac-b70b-b3cc35b1d7a1"). InnerVolumeSpecName "kube-api-access-tcj82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.525524 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcj82\" (UniqueName: \"kubernetes.io/projected/3959ab36-a688-40ac-b70b-b3cc35b1d7a1-kube-api-access-tcj82\") on node \"crc\" DevicePath \"\"" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.949817 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.950048 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.975148 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.985386 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563838-6fbsd" event={"ID":"3959ab36-a688-40ac-b70b-b3cc35b1d7a1","Type":"ContainerDied","Data":"767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d"} Mar 18 10:38:04 crc kubenswrapper[4733]: I0318 10:38:04.985461 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="767ef7bf1d3ba1aa0c4f7321a9fd4a583591dd67c6f6f39f86b5ad15f4ac2b8d" Mar 18 10:38:05 crc kubenswrapper[4733]: I0318 10:38:05.398400 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563832-njktw"] Mar 18 10:38:05 crc kubenswrapper[4733]: I0318 10:38:05.408699 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563832-njktw"] Mar 18 10:38:06 crc kubenswrapper[4733]: I0318 10:38:06.010021 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wxvvt" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="registry-server" probeResult="failure" output=< Mar 18 10:38:06 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:38:06 crc kubenswrapper[4733]: > Mar 18 10:38:07 crc kubenswrapper[4733]: I0318 10:38:07.189513 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac26b4cb-ac0b-4b78-9c5e-60c6563b478e" path="/var/lib/kubelet/pods/ac26b4cb-ac0b-4b78-9c5e-60c6563b478e/volumes" Mar 18 10:38:08 crc kubenswrapper[4733]: I0318 10:38:08.985972 4733 scope.go:117] "RemoveContainer" containerID="9bc7f39c4918c4a53f61ec2045418343aab6acedb5d7104271be60607764a8a9" Mar 18 10:38:09 crc kubenswrapper[4733]: I0318 10:38:09.013307 4733 scope.go:117] "RemoveContainer" containerID="fc8a98034f827fb8988cc2fa281e7a7c5e2bd32e772267e324591ed784c75b62" Mar 18 10:38:09 crc kubenswrapper[4733]: I0318 10:38:09.075507 4733 scope.go:117] "RemoveContainer" containerID="6a586c5fd4b77aeae152ada8c17b0c5946fd162e927825804e11d81559ba17f0" Mar 18 10:38:10 crc kubenswrapper[4733]: I0318 10:38:10.176415 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:38:11 crc kubenswrapper[4733]: I0318 10:38:11.033430 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c"} Mar 18 10:38:11 crc kubenswrapper[4733]: I0318 10:38:11.034334 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.026113 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.079306 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" exitCode=0 Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.079385 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c"} Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.079457 4733 scope.go:117] "RemoveContainer" containerID="b8c4d43890082484b9a1254cd9426c5cd83f45e4b1a61544da192f3cdebac71e" Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.080607 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:38:15 crc kubenswrapper[4733]: E0318 10:38:15.081108 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.115472 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:15 crc kubenswrapper[4733]: I0318 10:38:15.287038 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.093397 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wxvvt" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="registry-server" containerID="cri-o://a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa" gracePeriod=2 Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.176075 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:38:16 crc kubenswrapper[4733]: E0318 10:38:16.176466 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.660619 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.751956 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities\") pod \"b0c43d05-915d-4373-aae0-67a182acc4bc\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.752119 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content\") pod \"b0c43d05-915d-4373-aae0-67a182acc4bc\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.752215 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjpzr\" (UniqueName: \"kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr\") pod \"b0c43d05-915d-4373-aae0-67a182acc4bc\" (UID: \"b0c43d05-915d-4373-aae0-67a182acc4bc\") " Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.754146 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities" (OuterVolumeSpecName: "utilities") pod "b0c43d05-915d-4373-aae0-67a182acc4bc" (UID: "b0c43d05-915d-4373-aae0-67a182acc4bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.767658 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr" (OuterVolumeSpecName: "kube-api-access-xjpzr") pod "b0c43d05-915d-4373-aae0-67a182acc4bc" (UID: "b0c43d05-915d-4373-aae0-67a182acc4bc"). InnerVolumeSpecName "kube-api-access-xjpzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.853867 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.853901 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjpzr\" (UniqueName: \"kubernetes.io/projected/b0c43d05-915d-4373-aae0-67a182acc4bc-kube-api-access-xjpzr\") on node \"crc\" DevicePath \"\"" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.938745 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0c43d05-915d-4373-aae0-67a182acc4bc" (UID: "b0c43d05-915d-4373-aae0-67a182acc4bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:38:16 crc kubenswrapper[4733]: I0318 10:38:16.955709 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c43d05-915d-4373-aae0-67a182acc4bc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.126913 4733 generic.go:334] "Generic (PLEG): container finished" podID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerID="a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa" exitCode=0 Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.126974 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerDied","Data":"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa"} Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.127008 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxvvt" event={"ID":"b0c43d05-915d-4373-aae0-67a182acc4bc","Type":"ContainerDied","Data":"98f9eb05565700e8349c0932c41e77ad80e45cc728fe839b9e5a63014e0f9e85"} Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.127031 4733 scope.go:117] "RemoveContainer" containerID="a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.127073 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxvvt" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.161141 4733 scope.go:117] "RemoveContainer" containerID="c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.199551 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.200338 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wxvvt"] Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.207389 4733 scope.go:117] "RemoveContainer" containerID="7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.237794 4733 scope.go:117] "RemoveContainer" containerID="a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa" Mar 18 10:38:17 crc kubenswrapper[4733]: E0318 10:38:17.238998 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa\": container with ID starting with a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa not found: ID does not exist" containerID="a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.239069 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa"} err="failed to get container status \"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa\": rpc error: code = NotFound desc = could not find container \"a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa\": container with ID starting with a15d60390589ca2e549d7e30fc944c0099e4336af37991944c6539ec52e623aa not found: ID does not exist" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.239139 4733 scope.go:117] "RemoveContainer" containerID="c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74" Mar 18 10:38:17 crc kubenswrapper[4733]: E0318 10:38:17.240164 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74\": container with ID starting with c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74 not found: ID does not exist" containerID="c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.240227 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74"} err="failed to get container status \"c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74\": rpc error: code = NotFound desc = could not find container \"c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74\": container with ID starting with c49023eae46560082477167844c01c17c714b041e28ae5e5bfa596612d1eda74 not found: ID does not exist" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.240257 4733 scope.go:117] "RemoveContainer" containerID="7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e" Mar 18 10:38:17 crc kubenswrapper[4733]: E0318 10:38:17.240873 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e\": container with ID starting with 7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e not found: ID does not exist" containerID="7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e" Mar 18 10:38:17 crc kubenswrapper[4733]: I0318 10:38:17.241048 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e"} err="failed to get container status \"7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e\": rpc error: code = NotFound desc = could not find container \"7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e\": container with ID starting with 7261f0adf78b1ca0abdaa2369b3560c3c1bc5bc57577bda27c7f5f634951063e not found: ID does not exist" Mar 18 10:38:19 crc kubenswrapper[4733]: I0318 10:38:19.194583 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" path="/var/lib/kubelet/pods/b0c43d05-915d-4373-aae0-67a182acc4bc/volumes" Mar 18 10:38:30 crc kubenswrapper[4733]: I0318 10:38:30.176750 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:38:30 crc kubenswrapper[4733]: E0318 10:38:30.177882 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:38:30 crc kubenswrapper[4733]: I0318 10:38:30.177890 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:38:30 crc kubenswrapper[4733]: E0318 10:38:30.178464 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:38:41 crc kubenswrapper[4733]: I0318 10:38:41.185371 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:38:41 crc kubenswrapper[4733]: E0318 10:38:41.186577 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:38:43 crc kubenswrapper[4733]: I0318 10:38:43.571742 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:38:43 crc kubenswrapper[4733]: I0318 10:38:43.572238 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:38:44 crc kubenswrapper[4733]: I0318 10:38:44.176104 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:38:44 crc kubenswrapper[4733]: E0318 10:38:44.176661 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:38:56 crc kubenswrapper[4733]: I0318 10:38:56.175155 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:38:56 crc kubenswrapper[4733]: E0318 10:38:56.176813 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:38:56 crc kubenswrapper[4733]: I0318 10:38:56.176907 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:38:56 crc kubenswrapper[4733]: E0318 10:38:56.177138 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:39:07 crc kubenswrapper[4733]: I0318 10:39:07.176905 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:39:07 crc kubenswrapper[4733]: E0318 10:39:07.178328 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:39:09 crc kubenswrapper[4733]: I0318 10:39:09.220517 4733 scope.go:117] "RemoveContainer" containerID="da48f5028812280b5314f3d818c71b3049bdb0d8b1d5755bc74f1fedad4676d7" Mar 18 10:39:11 crc kubenswrapper[4733]: I0318 10:39:11.181379 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:39:11 crc kubenswrapper[4733]: E0318 10:39:11.181980 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:39:13 crc kubenswrapper[4733]: I0318 10:39:13.571700 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:39:13 crc kubenswrapper[4733]: I0318 10:39:13.572127 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:39:20 crc kubenswrapper[4733]: I0318 10:39:20.176215 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:39:20 crc kubenswrapper[4733]: E0318 10:39:20.177217 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:39:25 crc kubenswrapper[4733]: I0318 10:39:25.175961 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:39:25 crc kubenswrapper[4733]: E0318 10:39:25.176888 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:39:34 crc kubenswrapper[4733]: I0318 10:39:34.176649 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:39:34 crc kubenswrapper[4733]: E0318 10:39:34.177586 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:39:36 crc kubenswrapper[4733]: I0318 10:39:36.175634 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:39:36 crc kubenswrapper[4733]: E0318 10:39:36.176949 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:39:43 crc kubenswrapper[4733]: I0318 10:39:43.571798 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:39:43 crc kubenswrapper[4733]: I0318 10:39:43.572624 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:39:43 crc kubenswrapper[4733]: I0318 10:39:43.572691 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:39:43 crc kubenswrapper[4733]: I0318 10:39:43.573574 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:39:43 crc kubenswrapper[4733]: I0318 10:39:43.573670 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" gracePeriod=600 Mar 18 10:39:43 crc kubenswrapper[4733]: E0318 10:39:43.703727 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:39:44 crc kubenswrapper[4733]: I0318 10:39:44.024802 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" exitCode=0 Mar 18 10:39:44 crc kubenswrapper[4733]: I0318 10:39:44.024909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637"} Mar 18 10:39:44 crc kubenswrapper[4733]: I0318 10:39:44.025226 4733 scope.go:117] "RemoveContainer" containerID="18491327409d036c07217a5bf65332367e43c6f94559e59f3995caefe0f899d9" Mar 18 10:39:44 crc kubenswrapper[4733]: I0318 10:39:44.026060 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:39:44 crc kubenswrapper[4733]: E0318 10:39:44.026498 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:39:46 crc kubenswrapper[4733]: I0318 10:39:46.176800 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:39:46 crc kubenswrapper[4733]: E0318 10:39:46.177682 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:39:47 crc kubenswrapper[4733]: I0318 10:39:47.175864 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:39:47 crc kubenswrapper[4733]: E0318 10:39:47.176380 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:39:57 crc kubenswrapper[4733]: I0318 10:39:57.175984 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:39:57 crc kubenswrapper[4733]: E0318 10:39:57.176983 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.159937 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563840-lvjwz"] Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.160668 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="extract-utilities" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160681 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="extract-utilities" Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.160690 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="extract-content" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160696 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="extract-content" Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.160722 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="registry-server" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160727 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="registry-server" Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.160740 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3959ab36-a688-40ac-b70b-b3cc35b1d7a1" containerName="oc" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160746 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3959ab36-a688-40ac-b70b-b3cc35b1d7a1" containerName="oc" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160881 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3959ab36-a688-40ac-b70b-b3cc35b1d7a1" containerName="oc" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.160906 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c43d05-915d-4373-aae0-67a182acc4bc" containerName="registry-server" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.161466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.165120 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.165231 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.165871 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.175069 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.175446 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.176402 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:40:00 crc kubenswrapper[4733]: E0318 10:40:00.177060 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.177094 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563840-lvjwz"] Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.232749 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvhm5\" (UniqueName: \"kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5\") pod \"auto-csr-approver-29563840-lvjwz\" (UID: \"be59cc18-c769-443f-962f-042b8ba456b8\") " pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.334629 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvhm5\" (UniqueName: \"kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5\") pod \"auto-csr-approver-29563840-lvjwz\" (UID: \"be59cc18-c769-443f-962f-042b8ba456b8\") " pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.367623 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvhm5\" (UniqueName: \"kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5\") pod \"auto-csr-approver-29563840-lvjwz\" (UID: \"be59cc18-c769-443f-962f-042b8ba456b8\") " pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:00 crc kubenswrapper[4733]: I0318 10:40:00.496468 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:01 crc kubenswrapper[4733]: W0318 10:40:01.047646 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe59cc18_c769_443f_962f_042b8ba456b8.slice/crio-5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824 WatchSource:0}: Error finding container 5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824: Status 404 returned error can't find the container with id 5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824 Mar 18 10:40:01 crc kubenswrapper[4733]: I0318 10:40:01.049085 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563840-lvjwz"] Mar 18 10:40:01 crc kubenswrapper[4733]: I0318 10:40:01.193724 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" event={"ID":"be59cc18-c769-443f-962f-042b8ba456b8","Type":"ContainerStarted","Data":"5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824"} Mar 18 10:40:03 crc kubenswrapper[4733]: I0318 10:40:03.232679 4733 generic.go:334] "Generic (PLEG): container finished" podID="be59cc18-c769-443f-962f-042b8ba456b8" containerID="875056c9052fb9c1578503b6e7b5412bb4aa77c6a765ef7ede001dc3e0cb6698" exitCode=0 Mar 18 10:40:03 crc kubenswrapper[4733]: I0318 10:40:03.232962 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" event={"ID":"be59cc18-c769-443f-962f-042b8ba456b8","Type":"ContainerDied","Data":"875056c9052fb9c1578503b6e7b5412bb4aa77c6a765ef7ede001dc3e0cb6698"} Mar 18 10:40:04 crc kubenswrapper[4733]: I0318 10:40:04.641968 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:04 crc kubenswrapper[4733]: I0318 10:40:04.823357 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvhm5\" (UniqueName: \"kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5\") pod \"be59cc18-c769-443f-962f-042b8ba456b8\" (UID: \"be59cc18-c769-443f-962f-042b8ba456b8\") " Mar 18 10:40:04 crc kubenswrapper[4733]: I0318 10:40:04.832954 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5" (OuterVolumeSpecName: "kube-api-access-rvhm5") pod "be59cc18-c769-443f-962f-042b8ba456b8" (UID: "be59cc18-c769-443f-962f-042b8ba456b8"). InnerVolumeSpecName "kube-api-access-rvhm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:40:04 crc kubenswrapper[4733]: I0318 10:40:04.925589 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvhm5\" (UniqueName: \"kubernetes.io/projected/be59cc18-c769-443f-962f-042b8ba456b8-kube-api-access-rvhm5\") on node \"crc\" DevicePath \"\"" Mar 18 10:40:05 crc kubenswrapper[4733]: I0318 10:40:05.265167 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" event={"ID":"be59cc18-c769-443f-962f-042b8ba456b8","Type":"ContainerDied","Data":"5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824"} Mar 18 10:40:05 crc kubenswrapper[4733]: I0318 10:40:05.265258 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c4e2819f246915dad9f03c2c399d751618824fd626fb44a75fc69ec2dd1b824" Mar 18 10:40:05 crc kubenswrapper[4733]: I0318 10:40:05.265340 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563840-lvjwz" Mar 18 10:40:05 crc kubenswrapper[4733]: I0318 10:40:05.745117 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563834-7bqxq"] Mar 18 10:40:05 crc kubenswrapper[4733]: I0318 10:40:05.755499 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563834-7bqxq"] Mar 18 10:40:07 crc kubenswrapper[4733]: I0318 10:40:07.195676 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316" path="/var/lib/kubelet/pods/dc0f56eb-6c6a-49bf-9a12-ef5f2dd95316/volumes" Mar 18 10:40:09 crc kubenswrapper[4733]: I0318 10:40:09.599645 4733 scope.go:117] "RemoveContainer" containerID="2746f736c334d9ac3079e5dc9b5db5929c610a6933e47547e58536ec78e443c9" Mar 18 10:40:11 crc kubenswrapper[4733]: I0318 10:40:11.179826 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:40:11 crc kubenswrapper[4733]: E0318 10:40:11.180469 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:40:14 crc kubenswrapper[4733]: I0318 10:40:14.175978 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:40:14 crc kubenswrapper[4733]: I0318 10:40:14.176417 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:40:14 crc kubenswrapper[4733]: E0318 10:40:14.176689 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:40:14 crc kubenswrapper[4733]: E0318 10:40:14.176874 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:40:26 crc kubenswrapper[4733]: I0318 10:40:26.175559 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:40:26 crc kubenswrapper[4733]: E0318 10:40:26.176223 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:40:27 crc kubenswrapper[4733]: I0318 10:40:27.175585 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:40:27 crc kubenswrapper[4733]: E0318 10:40:27.176285 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:40:28 crc kubenswrapper[4733]: I0318 10:40:28.175925 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:40:28 crc kubenswrapper[4733]: E0318 10:40:28.176305 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:40:37 crc kubenswrapper[4733]: I0318 10:40:37.175333 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:40:37 crc kubenswrapper[4733]: E0318 10:40:37.176332 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:40:40 crc kubenswrapper[4733]: I0318 10:40:40.178461 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:40:40 crc kubenswrapper[4733]: E0318 10:40:40.179179 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:40:41 crc kubenswrapper[4733]: I0318 10:40:41.204033 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:40:41 crc kubenswrapper[4733]: E0318 10:40:41.205570 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:40:51 crc kubenswrapper[4733]: I0318 10:40:51.184513 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:40:51 crc kubenswrapper[4733]: E0318 10:40:51.185535 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:40:52 crc kubenswrapper[4733]: I0318 10:40:52.176662 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:40:52 crc kubenswrapper[4733]: E0318 10:40:52.177575 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:40:52 crc kubenswrapper[4733]: I0318 10:40:52.178318 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:40:52 crc kubenswrapper[4733]: E0318 10:40:52.178804 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:04 crc kubenswrapper[4733]: I0318 10:41:04.176242 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:41:04 crc kubenswrapper[4733]: I0318 10:41:04.177284 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:41:04 crc kubenswrapper[4733]: E0318 10:41:04.177620 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:41:04 crc kubenswrapper[4733]: E0318 10:41:04.177743 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:06 crc kubenswrapper[4733]: I0318 10:41:06.176281 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:41:06 crc kubenswrapper[4733]: E0318 10:41:06.177123 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:41:15 crc kubenswrapper[4733]: I0318 10:41:15.175935 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:41:15 crc kubenswrapper[4733]: E0318 10:41:15.177290 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:16 crc kubenswrapper[4733]: I0318 10:41:16.178357 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:41:16 crc kubenswrapper[4733]: E0318 10:41:16.179739 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:17.928263 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:17.928610 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:17.928690 4733 patch_prober.go:28] interesting pod/router-default-5444994796-xl5d7 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:17.928705 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-xl5d7" podUID="9c5f567e-b38f-44a0-b1fd-1a96857e811f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:17.973314 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" podUID="216f9239-7d2e-483e-a89f-0955a518aa4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.017921 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-767865f676-gkndg" podUID="216f9239-7d2e-483e-a89f-0955a518aa4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.177717 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.377438 4733 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2wc5m container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.422447 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podUID="cd9234ed-fcbc-4d81-9034-27d39b3df6ee" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.467448 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podUID="6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.512373 4733 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2wc5m container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.512421 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" podUID="9571ba80-f267-46ed-8d16-e44531cb0ce8" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.512493 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-sqr4g" podUID="cd9234ed-fcbc-4d81-9034-27d39b3df6ee" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.512561 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nskpj" podUID="6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:18.377515 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2wc5m" podUID="9571ba80-f267-46ed-8d16-e44531cb0ce8" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.143653 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-dr9dg" podUID="03476444-8ff8-4b1e-bcbc-ee654241370b" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.52:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.143788 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z6qb2 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.143821 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" podUID="8ae3847e-6357-46a1-9578-88deb6e1531b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.144284 4733 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-kd6gw container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.144324 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kd6gw" podUID="9b0edb65-3bcf-484f-9707-d8124df1ec88" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: E0318 10:41:19.154965 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.168666 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-pc5zz" podUID="4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.168760 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z6qb2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.168783 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z6qb2" podUID="8ae3847e-6357-46a1-9578-88deb6e1531b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.168926 4733 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 10:41:19 crc kubenswrapper[4733]: I0318 10:41:19.168947 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 10:41:27 crc kubenswrapper[4733]: I0318 10:41:27.175581 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:41:27 crc kubenswrapper[4733]: E0318 10:41:27.176521 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:41:29 crc kubenswrapper[4733]: I0318 10:41:29.175570 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:41:29 crc kubenswrapper[4733]: E0318 10:41:29.176293 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:30 crc kubenswrapper[4733]: I0318 10:41:30.176109 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:41:30 crc kubenswrapper[4733]: E0318 10:41:30.177720 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:41:38 crc kubenswrapper[4733]: I0318 10:41:38.176243 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:41:38 crc kubenswrapper[4733]: E0318 10:41:38.177724 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.072408 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mr65v"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.081565 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0937-account-create-update-bfx7n"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.088867 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5gwmb"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.095842 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5e1e-account-create-update-r9bb4"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.103370 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mr65v"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.109323 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5e1e-account-create-update-r9bb4"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.113994 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0937-account-create-update-bfx7n"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.118619 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5gwmb"] Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.185074 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07730b47-54ba-4b79-952e-6fb12b3b5279" path="/var/lib/kubelet/pods/07730b47-54ba-4b79-952e-6fb12b3b5279/volumes" Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.185774 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5425fb-7059-4262-9c68-1420a5f3b4f1" path="/var/lib/kubelet/pods/0a5425fb-7059-4262-9c68-1420a5f3b4f1/volumes" Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.186327 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a7c351-0be1-4547-bacc-8ff02cb59328" path="/var/lib/kubelet/pods/30a7c351-0be1-4547-bacc-8ff02cb59328/volumes" Mar 18 10:41:41 crc kubenswrapper[4733]: I0318 10:41:41.186858 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6346115-9e7a-4489-916d-a129aa83a6dd" path="/var/lib/kubelet/pods/a6346115-9e7a-4489-916d-a129aa83a6dd/volumes" Mar 18 10:41:42 crc kubenswrapper[4733]: I0318 10:41:42.176047 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:41:42 crc kubenswrapper[4733]: E0318 10:41:42.176553 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:41:42 crc kubenswrapper[4733]: I0318 10:41:42.177023 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:41:42 crc kubenswrapper[4733]: E0318 10:41:42.177438 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.045091 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5795-account-create-update-nkww7"] Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.057833 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-fvlqt"] Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.064256 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-fvlqt"] Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.070994 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5795-account-create-update-nkww7"] Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.187057 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d4401f-2343-41fa-82ae-877674337bf4" path="/var/lib/kubelet/pods/84d4401f-2343-41fa-82ae-877674337bf4/volumes" Mar 18 10:41:45 crc kubenswrapper[4733]: I0318 10:41:45.187884 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5315db-fb68-4558-85c1-cf538d0e2770" path="/var/lib/kubelet/pods/ff5315db-fb68-4558-85c1-cf538d0e2770/volumes" Mar 18 10:41:53 crc kubenswrapper[4733]: I0318 10:41:53.175841 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:41:53 crc kubenswrapper[4733]: I0318 10:41:53.176644 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:41:53 crc kubenswrapper[4733]: E0318 10:41:53.176989 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:41:53 crc kubenswrapper[4733]: E0318 10:41:53.177020 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.706035 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:41:55 crc kubenswrapper[4733]: E0318 10:41:55.725603 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be59cc18-c769-443f-962f-042b8ba456b8" containerName="oc" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.725648 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="be59cc18-c769-443f-962f-042b8ba456b8" containerName="oc" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.726262 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="be59cc18-c769-443f-962f-042b8ba456b8" containerName="oc" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.729069 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.729263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.772029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.772160 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.772221 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5462\" (UniqueName: \"kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.873802 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.873851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5462\" (UniqueName: \"kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.873903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.874459 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.874461 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:55 crc kubenswrapper[4733]: I0318 10:41:55.899624 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5462\" (UniqueName: \"kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462\") pod \"certified-operators-tljzm\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:56 crc kubenswrapper[4733]: I0318 10:41:56.083429 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:41:56 crc kubenswrapper[4733]: I0318 10:41:56.175991 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:41:56 crc kubenswrapper[4733]: E0318 10:41:56.176651 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:41:56 crc kubenswrapper[4733]: W0318 10:41:56.611357 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod137e4764_e111_4ef3_ac40_3f8a07e4df8a.slice/crio-387eba0329a2b04d4a916eb64b70411eb77b955884ab7529c7638dae04a05a85 WatchSource:0}: Error finding container 387eba0329a2b04d4a916eb64b70411eb77b955884ab7529c7638dae04a05a85: Status 404 returned error can't find the container with id 387eba0329a2b04d4a916eb64b70411eb77b955884ab7529c7638dae04a05a85 Mar 18 10:41:56 crc kubenswrapper[4733]: I0318 10:41:56.623052 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:41:57 crc kubenswrapper[4733]: I0318 10:41:57.551079 4733 generic.go:334] "Generic (PLEG): container finished" podID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerID="4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307" exitCode=0 Mar 18 10:41:57 crc kubenswrapper[4733]: I0318 10:41:57.551224 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerDied","Data":"4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307"} Mar 18 10:41:57 crc kubenswrapper[4733]: I0318 10:41:57.551639 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerStarted","Data":"387eba0329a2b04d4a916eb64b70411eb77b955884ab7529c7638dae04a05a85"} Mar 18 10:41:59 crc kubenswrapper[4733]: I0318 10:41:59.572332 4733 generic.go:334] "Generic (PLEG): container finished" podID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerID="fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8" exitCode=0 Mar 18 10:41:59 crc kubenswrapper[4733]: I0318 10:41:59.572451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerDied","Data":"fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8"} Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.158521 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563842-78hdh"] Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.160243 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.170670 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563842-78hdh"] Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.187837 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.187894 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.188163 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.357985 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfnb8\" (UniqueName: \"kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8\") pod \"auto-csr-approver-29563842-78hdh\" (UID: \"8a6751ff-c17a-472f-b315-06edff233f07\") " pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.459407 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfnb8\" (UniqueName: \"kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8\") pod \"auto-csr-approver-29563842-78hdh\" (UID: \"8a6751ff-c17a-472f-b315-06edff233f07\") " pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.498430 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfnb8\" (UniqueName: \"kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8\") pod \"auto-csr-approver-29563842-78hdh\" (UID: \"8a6751ff-c17a-472f-b315-06edff233f07\") " pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.524834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.585105 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerStarted","Data":"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b"} Mar 18 10:42:00 crc kubenswrapper[4733]: I0318 10:42:00.628105 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tljzm" podStartSLOduration=3.1727671 podStartE2EDuration="5.628085808s" podCreationTimestamp="2026-03-18 10:41:55 +0000 UTC" firstStartedPulling="2026-03-18 10:41:57.554577155 +0000 UTC m=+1757.046311520" lastFinishedPulling="2026-03-18 10:42:00.009895863 +0000 UTC m=+1759.501630228" observedRunningTime="2026-03-18 10:42:00.624625411 +0000 UTC m=+1760.116359806" watchObservedRunningTime="2026-03-18 10:42:00.628085808 +0000 UTC m=+1760.119820143" Mar 18 10:42:01 crc kubenswrapper[4733]: I0318 10:42:01.011990 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563842-78hdh"] Mar 18 10:42:01 crc kubenswrapper[4733]: W0318 10:42:01.027377 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a6751ff_c17a_472f_b315_06edff233f07.slice/crio-c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5 WatchSource:0}: Error finding container c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5: Status 404 returned error can't find the container with id c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5 Mar 18 10:42:01 crc kubenswrapper[4733]: I0318 10:42:01.594468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563842-78hdh" event={"ID":"8a6751ff-c17a-472f-b315-06edff233f07","Type":"ContainerStarted","Data":"c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5"} Mar 18 10:42:02 crc kubenswrapper[4733]: I0318 10:42:02.606532 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563842-78hdh" event={"ID":"8a6751ff-c17a-472f-b315-06edff233f07","Type":"ContainerStarted","Data":"56204ccc444d383e182644d7909ef75775c83c4da5e6940b6afcaf6c25fa0fc2"} Mar 18 10:42:02 crc kubenswrapper[4733]: I0318 10:42:02.628997 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563842-78hdh" podStartSLOduration=1.483118036 podStartE2EDuration="2.628976965s" podCreationTimestamp="2026-03-18 10:42:00 +0000 UTC" firstStartedPulling="2026-03-18 10:42:01.031261796 +0000 UTC m=+1760.522996161" lastFinishedPulling="2026-03-18 10:42:02.177120725 +0000 UTC m=+1761.668855090" observedRunningTime="2026-03-18 10:42:02.621952467 +0000 UTC m=+1762.113686872" watchObservedRunningTime="2026-03-18 10:42:02.628976965 +0000 UTC m=+1762.120711290" Mar 18 10:42:03 crc kubenswrapper[4733]: I0318 10:42:03.621802 4733 generic.go:334] "Generic (PLEG): container finished" podID="8a6751ff-c17a-472f-b315-06edff233f07" containerID="56204ccc444d383e182644d7909ef75775c83c4da5e6940b6afcaf6c25fa0fc2" exitCode=0 Mar 18 10:42:03 crc kubenswrapper[4733]: I0318 10:42:03.621898 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563842-78hdh" event={"ID":"8a6751ff-c17a-472f-b315-06edff233f07","Type":"ContainerDied","Data":"56204ccc444d383e182644d7909ef75775c83c4da5e6940b6afcaf6c25fa0fc2"} Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.055585 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.244811 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfnb8\" (UniqueName: \"kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8\") pod \"8a6751ff-c17a-472f-b315-06edff233f07\" (UID: \"8a6751ff-c17a-472f-b315-06edff233f07\") " Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.253151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8" (OuterVolumeSpecName: "kube-api-access-hfnb8") pod "8a6751ff-c17a-472f-b315-06edff233f07" (UID: "8a6751ff-c17a-472f-b315-06edff233f07"). InnerVolumeSpecName "kube-api-access-hfnb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.347100 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfnb8\" (UniqueName: \"kubernetes.io/projected/8a6751ff-c17a-472f-b315-06edff233f07-kube-api-access-hfnb8\") on node \"crc\" DevicePath \"\"" Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.644914 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563842-78hdh" event={"ID":"8a6751ff-c17a-472f-b315-06edff233f07","Type":"ContainerDied","Data":"c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5"} Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.644975 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8194ce2558598ec34fe5d174c0be4529d1eed31838e848e03a5a42254a157e5" Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.645063 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563842-78hdh" Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.707379 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563836-4x58h"] Mar 18 10:42:05 crc kubenswrapper[4733]: I0318 10:42:05.719313 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563836-4x58h"] Mar 18 10:42:06 crc kubenswrapper[4733]: I0318 10:42:06.083914 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:06 crc kubenswrapper[4733]: I0318 10:42:06.083972 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:06 crc kubenswrapper[4733]: I0318 10:42:06.165234 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:06 crc kubenswrapper[4733]: I0318 10:42:06.746665 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:06 crc kubenswrapper[4733]: I0318 10:42:06.814500 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:42:07 crc kubenswrapper[4733]: I0318 10:42:07.176215 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:42:07 crc kubenswrapper[4733]: E0318 10:42:07.176629 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:42:07 crc kubenswrapper[4733]: I0318 10:42:07.193908 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d31d006c-81a6-4bbb-a44a-fda966944372" path="/var/lib/kubelet/pods/d31d006c-81a6-4bbb-a44a-fda966944372/volumes" Mar 18 10:42:08 crc kubenswrapper[4733]: I0318 10:42:08.177971 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:42:08 crc kubenswrapper[4733]: E0318 10:42:08.178333 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:42:08 crc kubenswrapper[4733]: I0318 10:42:08.682899 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tljzm" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="registry-server" containerID="cri-o://5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b" gracePeriod=2 Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.175630 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:42:09 crc kubenswrapper[4733]: E0318 10:42:09.176936 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.243474 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.320075 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities\") pod \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.320180 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5462\" (UniqueName: \"kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462\") pod \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.320393 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content\") pod \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\" (UID: \"137e4764-e111-4ef3-ac40-3f8a07e4df8a\") " Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.322223 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities" (OuterVolumeSpecName: "utilities") pod "137e4764-e111-4ef3-ac40-3f8a07e4df8a" (UID: "137e4764-e111-4ef3-ac40-3f8a07e4df8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.331317 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462" (OuterVolumeSpecName: "kube-api-access-t5462") pod "137e4764-e111-4ef3-ac40-3f8a07e4df8a" (UID: "137e4764-e111-4ef3-ac40-3f8a07e4df8a"). InnerVolumeSpecName "kube-api-access-t5462". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.407281 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "137e4764-e111-4ef3-ac40-3f8a07e4df8a" (UID: "137e4764-e111-4ef3-ac40-3f8a07e4df8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.422122 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.422165 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/137e4764-e111-4ef3-ac40-3f8a07e4df8a-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.422206 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5462\" (UniqueName: \"kubernetes.io/projected/137e4764-e111-4ef3-ac40-3f8a07e4df8a-kube-api-access-t5462\") on node \"crc\" DevicePath \"\"" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.695072 4733 generic.go:334] "Generic (PLEG): container finished" podID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerID="5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b" exitCode=0 Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.695136 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tljzm" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.695157 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerDied","Data":"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b"} Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.695633 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tljzm" event={"ID":"137e4764-e111-4ef3-ac40-3f8a07e4df8a","Type":"ContainerDied","Data":"387eba0329a2b04d4a916eb64b70411eb77b955884ab7529c7638dae04a05a85"} Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.695705 4733 scope.go:117] "RemoveContainer" containerID="5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.711839 4733 scope.go:117] "RemoveContainer" containerID="ce9a99c6df86d54aacd4034e75a79275a1f1a3fe6a26a1b9d309967e3b0b146b" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.725286 4733 scope.go:117] "RemoveContainer" containerID="fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.752579 4733 scope.go:117] "RemoveContainer" containerID="754b489534a1d2e07cfe28b803b1041b8c35b7a8d870ab7643873669d480405d" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.761408 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.773531 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tljzm"] Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.806548 4733 scope.go:117] "RemoveContainer" containerID="1a52840f130018d9dd9a4d4957090d0bfe7cddccea8c86d998fc7ce63f88d2c3" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.821803 4733 scope.go:117] "RemoveContainer" containerID="4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.879893 4733 scope.go:117] "RemoveContainer" containerID="5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b" Mar 18 10:42:09 crc kubenswrapper[4733]: E0318 10:42:09.880634 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b\": container with ID starting with 5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b not found: ID does not exist" containerID="5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.880687 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b"} err="failed to get container status \"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b\": rpc error: code = NotFound desc = could not find container \"5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b\": container with ID starting with 5d36f5790f642401acb40728675a784e6d4f0ef6401d704d6528ca8351f6c46b not found: ID does not exist" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.880723 4733 scope.go:117] "RemoveContainer" containerID="fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8" Mar 18 10:42:09 crc kubenswrapper[4733]: E0318 10:42:09.881162 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8\": container with ID starting with fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8 not found: ID does not exist" containerID="fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.881238 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8"} err="failed to get container status \"fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8\": rpc error: code = NotFound desc = could not find container \"fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8\": container with ID starting with fb254defcfea9fc1997376ee41afbe7b64c90aee6f3e5d2915155a5e3c51f9c8 not found: ID does not exist" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.881266 4733 scope.go:117] "RemoveContainer" containerID="4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307" Mar 18 10:42:09 crc kubenswrapper[4733]: E0318 10:42:09.881741 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307\": container with ID starting with 4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307 not found: ID does not exist" containerID="4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.881787 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307"} err="failed to get container status \"4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307\": rpc error: code = NotFound desc = could not find container \"4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307\": container with ID starting with 4fcc3926dcfa6f9c4be1799c42751eb86e56709f553565e31fa0e718a378e307 not found: ID does not exist" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.901607 4733 scope.go:117] "RemoveContainer" containerID="cb5c331f367d49d9d35cab0a581b0fd4e3d8921934861b35f887d6648ae09cfb" Mar 18 10:42:09 crc kubenswrapper[4733]: I0318 10:42:09.974973 4733 scope.go:117] "RemoveContainer" containerID="c39830f7afc41d4e539449a18ea110efdcfaa942a6d99809d36a31233d6cb82b" Mar 18 10:42:10 crc kubenswrapper[4733]: I0318 10:42:10.009762 4733 scope.go:117] "RemoveContainer" containerID="4d1446897edd6664fd044255842a94c2933d1bab0fe0d09f54123f9a53833063" Mar 18 10:42:10 crc kubenswrapper[4733]: I0318 10:42:10.077118 4733 scope.go:117] "RemoveContainer" containerID="881999bdad04a088176edfb2a1165638bbb818ce5892ed189c2612e4735ca703" Mar 18 10:42:11 crc kubenswrapper[4733]: I0318 10:42:11.188221 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" path="/var/lib/kubelet/pods/137e4764-e111-4ef3-ac40-3f8a07e4df8a/volumes" Mar 18 10:42:12 crc kubenswrapper[4733]: I0318 10:42:12.048743 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-xb87f"] Mar 18 10:42:12 crc kubenswrapper[4733]: I0318 10:42:12.057876 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-xb87f"] Mar 18 10:42:13 crc kubenswrapper[4733]: I0318 10:42:13.184643 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d533f566-eded-44ca-b276-7e3d437f9fce" path="/var/lib/kubelet/pods/d533f566-eded-44ca-b276-7e3d437f9fce/volumes" Mar 18 10:42:22 crc kubenswrapper[4733]: I0318 10:42:22.176117 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:42:22 crc kubenswrapper[4733]: I0318 10:42:22.176679 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:42:22 crc kubenswrapper[4733]: I0318 10:42:22.176856 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:42:22 crc kubenswrapper[4733]: E0318 10:42:22.176894 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:42:22 crc kubenswrapper[4733]: E0318 10:42:22.177032 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:42:22 crc kubenswrapper[4733]: E0318 10:42:22.177386 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:42:29 crc kubenswrapper[4733]: I0318 10:42:29.050719 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ptbmt"] Mar 18 10:42:29 crc kubenswrapper[4733]: I0318 10:42:29.064117 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ptbmt"] Mar 18 10:42:29 crc kubenswrapper[4733]: I0318 10:42:29.190946 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c8f7bc-4162-4903-b3f9-96c8736a27b8" path="/var/lib/kubelet/pods/63c8f7bc-4162-4903-b3f9-96c8736a27b8/volumes" Mar 18 10:42:33 crc kubenswrapper[4733]: I0318 10:42:33.176564 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:42:33 crc kubenswrapper[4733]: E0318 10:42:33.177529 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:42:35 crc kubenswrapper[4733]: I0318 10:42:35.175923 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:42:35 crc kubenswrapper[4733]: E0318 10:42:35.177019 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:42:36 crc kubenswrapper[4733]: I0318 10:42:36.176457 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:42:36 crc kubenswrapper[4733]: E0318 10:42:36.177266 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:42:45 crc kubenswrapper[4733]: I0318 10:42:45.176281 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:42:45 crc kubenswrapper[4733]: E0318 10:42:45.177539 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:42:46 crc kubenswrapper[4733]: I0318 10:42:46.176278 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:42:46 crc kubenswrapper[4733]: E0318 10:42:46.176731 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:42:49 crc kubenswrapper[4733]: I0318 10:42:49.178424 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:42:49 crc kubenswrapper[4733]: E0318 10:42:49.178792 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:42:56 crc kubenswrapper[4733]: I0318 10:42:56.175399 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:42:56 crc kubenswrapper[4733]: E0318 10:42:56.176083 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:43:01 crc kubenswrapper[4733]: I0318 10:43:01.180978 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:43:01 crc kubenswrapper[4733]: E0318 10:43:01.181535 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:43:01 crc kubenswrapper[4733]: I0318 10:43:01.181670 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:43:01 crc kubenswrapper[4733]: E0318 10:43:01.182056 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:43:07 crc kubenswrapper[4733]: I0318 10:43:07.175495 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:43:07 crc kubenswrapper[4733]: E0318 10:43:07.176344 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:43:10 crc kubenswrapper[4733]: I0318 10:43:10.201844 4733 scope.go:117] "RemoveContainer" containerID="5eade19f29d1bfd378a24e80adf648a30d27b707a4c07763d5b7990ffd79ce55" Mar 18 10:43:10 crc kubenswrapper[4733]: I0318 10:43:10.252105 4733 scope.go:117] "RemoveContainer" containerID="7076d89bfeedd95679091846270edbde667d954c3e6fdb8ee00f499b50144915" Mar 18 10:43:12 crc kubenswrapper[4733]: I0318 10:43:12.176422 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:43:12 crc kubenswrapper[4733]: E0318 10:43:12.177163 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:43:13 crc kubenswrapper[4733]: I0318 10:43:13.177003 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:43:14 crc kubenswrapper[4733]: I0318 10:43:14.320589 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98"} Mar 18 10:43:14 crc kubenswrapper[4733]: I0318 10:43:14.321740 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:43:18 crc kubenswrapper[4733]: I0318 10:43:18.362606 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" exitCode=0 Mar 18 10:43:18 crc kubenswrapper[4733]: I0318 10:43:18.362742 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98"} Mar 18 10:43:18 crc kubenswrapper[4733]: I0318 10:43:18.362973 4733 scope.go:117] "RemoveContainer" containerID="39f4b0f268d8d6f6613db69085678023c98d9ef75625187f285b2a88548a855a" Mar 18 10:43:18 crc kubenswrapper[4733]: I0318 10:43:18.364016 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:43:18 crc kubenswrapper[4733]: E0318 10:43:18.364691 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:43:20 crc kubenswrapper[4733]: I0318 10:43:20.175867 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:43:20 crc kubenswrapper[4733]: E0318 10:43:20.176488 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:43:23 crc kubenswrapper[4733]: I0318 10:43:23.177515 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:43:23 crc kubenswrapper[4733]: I0318 10:43:23.417077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8"} Mar 18 10:43:23 crc kubenswrapper[4733]: I0318 10:43:23.418017 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:43:28 crc kubenswrapper[4733]: I0318 10:43:28.470077 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" exitCode=0 Mar 18 10:43:28 crc kubenswrapper[4733]: I0318 10:43:28.470272 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8"} Mar 18 10:43:28 crc kubenswrapper[4733]: I0318 10:43:28.471156 4733 scope.go:117] "RemoveContainer" containerID="34bcdecb0fa459a1d6253fb10ac79d88f89fe0809b9e5dec1a6ca99fb8f7810c" Mar 18 10:43:28 crc kubenswrapper[4733]: I0318 10:43:28.472253 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:43:28 crc kubenswrapper[4733]: E0318 10:43:28.472815 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:43:32 crc kubenswrapper[4733]: I0318 10:43:32.177100 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:43:32 crc kubenswrapper[4733]: E0318 10:43:32.179575 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:43:34 crc kubenswrapper[4733]: I0318 10:43:34.175798 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:43:34 crc kubenswrapper[4733]: E0318 10:43:34.176460 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:43:43 crc kubenswrapper[4733]: I0318 10:43:43.177241 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:43:43 crc kubenswrapper[4733]: E0318 10:43:43.178048 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:43:47 crc kubenswrapper[4733]: I0318 10:43:47.175676 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:43:47 crc kubenswrapper[4733]: E0318 10:43:47.176685 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:43:49 crc kubenswrapper[4733]: I0318 10:43:49.176290 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:43:49 crc kubenswrapper[4733]: E0318 10:43:49.177000 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:43:54 crc kubenswrapper[4733]: I0318 10:43:54.176175 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:43:54 crc kubenswrapper[4733]: E0318 10:43:54.177115 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:43:58 crc kubenswrapper[4733]: I0318 10:43:58.176098 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:43:58 crc kubenswrapper[4733]: E0318 10:43:58.177037 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.148336 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563844-cdw7c"] Mar 18 10:44:00 crc kubenswrapper[4733]: E0318 10:44:00.148947 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="extract-content" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.148958 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="extract-content" Mar 18 10:44:00 crc kubenswrapper[4733]: E0318 10:44:00.148967 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6751ff-c17a-472f-b315-06edff233f07" containerName="oc" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.148973 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6751ff-c17a-472f-b315-06edff233f07" containerName="oc" Mar 18 10:44:00 crc kubenswrapper[4733]: E0318 10:44:00.148991 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="registry-server" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.148997 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="registry-server" Mar 18 10:44:00 crc kubenswrapper[4733]: E0318 10:44:00.149005 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="extract-utilities" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.149011 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="extract-utilities" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.149154 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="137e4764-e111-4ef3-ac40-3f8a07e4df8a" containerName="registry-server" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.149178 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6751ff-c17a-472f-b315-06edff233f07" containerName="oc" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.149666 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.151943 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.152237 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.152480 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.154660 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563844-cdw7c"] Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.328223 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw8tl\" (UniqueName: \"kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl\") pod \"auto-csr-approver-29563844-cdw7c\" (UID: \"a764647f-75c8-4ce3-82fb-b2b729a895a0\") " pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.429876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw8tl\" (UniqueName: \"kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl\") pod \"auto-csr-approver-29563844-cdw7c\" (UID: \"a764647f-75c8-4ce3-82fb-b2b729a895a0\") " pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.468484 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw8tl\" (UniqueName: \"kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl\") pod \"auto-csr-approver-29563844-cdw7c\" (UID: \"a764647f-75c8-4ce3-82fb-b2b729a895a0\") " pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.515854 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.784632 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563844-cdw7c"] Mar 18 10:44:00 crc kubenswrapper[4733]: I0318 10:44:00.804582 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:44:01 crc kubenswrapper[4733]: I0318 10:44:01.063831 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" event={"ID":"a764647f-75c8-4ce3-82fb-b2b729a895a0","Type":"ContainerStarted","Data":"d7f8967a5b8e446f7586f68c0936471f337551cf637a1d1b3e484af1a9c36311"} Mar 18 10:44:03 crc kubenswrapper[4733]: I0318 10:44:03.080407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" event={"ID":"a764647f-75c8-4ce3-82fb-b2b729a895a0","Type":"ContainerStarted","Data":"4e31a574451b3bae3c66b0663fefb42d9a6b941c8dd7f0cf1a6c603f449c0e3b"} Mar 18 10:44:03 crc kubenswrapper[4733]: I0318 10:44:03.107211 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" podStartSLOduration=1.302097448 podStartE2EDuration="3.107181679s" podCreationTimestamp="2026-03-18 10:44:00 +0000 UTC" firstStartedPulling="2026-03-18 10:44:00.803921483 +0000 UTC m=+1880.295655848" lastFinishedPulling="2026-03-18 10:44:02.609005714 +0000 UTC m=+1882.100740079" observedRunningTime="2026-03-18 10:44:03.101324904 +0000 UTC m=+1882.593059259" watchObservedRunningTime="2026-03-18 10:44:03.107181679 +0000 UTC m=+1882.598916004" Mar 18 10:44:03 crc kubenswrapper[4733]: I0318 10:44:03.176340 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:44:03 crc kubenswrapper[4733]: E0318 10:44:03.176588 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:44:04 crc kubenswrapper[4733]: I0318 10:44:04.091343 4733 generic.go:334] "Generic (PLEG): container finished" podID="a764647f-75c8-4ce3-82fb-b2b729a895a0" containerID="4e31a574451b3bae3c66b0663fefb42d9a6b941c8dd7f0cf1a6c603f449c0e3b" exitCode=0 Mar 18 10:44:04 crc kubenswrapper[4733]: I0318 10:44:04.091404 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" event={"ID":"a764647f-75c8-4ce3-82fb-b2b729a895a0","Type":"ContainerDied","Data":"4e31a574451b3bae3c66b0663fefb42d9a6b941c8dd7f0cf1a6c603f449c0e3b"} Mar 18 10:44:06 crc kubenswrapper[4733]: I0318 10:44:06.418640 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:06 crc kubenswrapper[4733]: I0318 10:44:06.561410 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw8tl\" (UniqueName: \"kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl\") pod \"a764647f-75c8-4ce3-82fb-b2b729a895a0\" (UID: \"a764647f-75c8-4ce3-82fb-b2b729a895a0\") " Mar 18 10:44:06 crc kubenswrapper[4733]: I0318 10:44:06.579546 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl" (OuterVolumeSpecName: "kube-api-access-dw8tl") pod "a764647f-75c8-4ce3-82fb-b2b729a895a0" (UID: "a764647f-75c8-4ce3-82fb-b2b729a895a0"). InnerVolumeSpecName "kube-api-access-dw8tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:44:06 crc kubenswrapper[4733]: I0318 10:44:06.663124 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw8tl\" (UniqueName: \"kubernetes.io/projected/a764647f-75c8-4ce3-82fb-b2b729a895a0-kube-api-access-dw8tl\") on node \"crc\" DevicePath \"\"" Mar 18 10:44:07 crc kubenswrapper[4733]: I0318 10:44:07.181933 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" Mar 18 10:44:07 crc kubenswrapper[4733]: I0318 10:44:07.196335 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563844-cdw7c" event={"ID":"a764647f-75c8-4ce3-82fb-b2b729a895a0","Type":"ContainerDied","Data":"d7f8967a5b8e446f7586f68c0936471f337551cf637a1d1b3e484af1a9c36311"} Mar 18 10:44:07 crc kubenswrapper[4733]: I0318 10:44:07.196395 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f8967a5b8e446f7586f68c0936471f337551cf637a1d1b3e484af1a9c36311" Mar 18 10:44:07 crc kubenswrapper[4733]: I0318 10:44:07.524702 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563838-6fbsd"] Mar 18 10:44:07 crc kubenswrapper[4733]: I0318 10:44:07.535259 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563838-6fbsd"] Mar 18 10:44:09 crc kubenswrapper[4733]: I0318 10:44:09.175471 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:44:09 crc kubenswrapper[4733]: E0318 10:44:09.176605 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:44:09 crc kubenswrapper[4733]: I0318 10:44:09.194840 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3959ab36-a688-40ac-b70b-b3cc35b1d7a1" path="/var/lib/kubelet/pods/3959ab36-a688-40ac-b70b-b3cc35b1d7a1/volumes" Mar 18 10:44:10 crc kubenswrapper[4733]: I0318 10:44:10.177732 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:44:10 crc kubenswrapper[4733]: E0318 10:44:10.178116 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:44:10 crc kubenswrapper[4733]: I0318 10:44:10.327800 4733 scope.go:117] "RemoveContainer" containerID="ac18342a3539a4f4eb0b18430ab3c33bd2af4e21dfc3695dc34573a145ad949d" Mar 18 10:44:18 crc kubenswrapper[4733]: I0318 10:44:18.175468 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:44:18 crc kubenswrapper[4733]: E0318 10:44:18.176322 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:44:21 crc kubenswrapper[4733]: I0318 10:44:21.188725 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:44:21 crc kubenswrapper[4733]: E0318 10:44:21.189445 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:44:23 crc kubenswrapper[4733]: I0318 10:44:23.175412 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:44:23 crc kubenswrapper[4733]: E0318 10:44:23.175885 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:44:30 crc kubenswrapper[4733]: I0318 10:44:30.176069 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:44:30 crc kubenswrapper[4733]: E0318 10:44:30.177219 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:44:35 crc kubenswrapper[4733]: I0318 10:44:35.176306 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:44:35 crc kubenswrapper[4733]: E0318 10:44:35.177043 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:44:36 crc kubenswrapper[4733]: I0318 10:44:36.178945 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:44:36 crc kubenswrapper[4733]: E0318 10:44:36.179385 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:44:42 crc kubenswrapper[4733]: I0318 10:44:42.176962 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:44:42 crc kubenswrapper[4733]: E0318 10:44:42.178223 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:44:46 crc kubenswrapper[4733]: I0318 10:44:46.175622 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:44:46 crc kubenswrapper[4733]: I0318 10:44:46.559098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337"} Mar 18 10:44:50 crc kubenswrapper[4733]: I0318 10:44:50.175567 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:44:50 crc kubenswrapper[4733]: E0318 10:44:50.176671 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:44:53 crc kubenswrapper[4733]: I0318 10:44:53.175252 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:44:53 crc kubenswrapper[4733]: E0318 10:44:53.176510 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.161104 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj"] Mar 18 10:45:00 crc kubenswrapper[4733]: E0318 10:45:00.164033 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a764647f-75c8-4ce3-82fb-b2b729a895a0" containerName="oc" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.164225 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a764647f-75c8-4ce3-82fb-b2b729a895a0" containerName="oc" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.164655 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a764647f-75c8-4ce3-82fb-b2b729a895a0" containerName="oc" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.166361 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.171716 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.174142 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj"] Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.175308 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.231379 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.231465 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4khzz\" (UniqueName: \"kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.231545 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.334074 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.334131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4khzz\" (UniqueName: \"kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.334185 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.335220 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.344289 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.361756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4khzz\" (UniqueName: \"kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz\") pod \"collect-profiles-29563845-8g7fj\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:00 crc kubenswrapper[4733]: I0318 10:45:00.488824 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:01 crc kubenswrapper[4733]: W0318 10:45:01.003377 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb43c9f9c_7dd6_45fd_b0e2_66e4e6de9852.slice/crio-7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f WatchSource:0}: Error finding container 7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f: Status 404 returned error can't find the container with id 7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f Mar 18 10:45:01 crc kubenswrapper[4733]: I0318 10:45:01.004942 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj"] Mar 18 10:45:01 crc kubenswrapper[4733]: I0318 10:45:01.026455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" event={"ID":"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852","Type":"ContainerStarted","Data":"7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f"} Mar 18 10:45:01 crc kubenswrapper[4733]: I0318 10:45:01.179946 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:01 crc kubenswrapper[4733]: E0318 10:45:01.180599 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:02 crc kubenswrapper[4733]: I0318 10:45:02.428951 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" event={"ID":"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852","Type":"ContainerDied","Data":"10b0b28bf2347978c5f83c64e4ba31d03fc0ca97939b344013089211d7ff71e7"} Mar 18 10:45:02 crc kubenswrapper[4733]: I0318 10:45:02.428642 4733 generic.go:334] "Generic (PLEG): container finished" podID="b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" containerID="10b0b28bf2347978c5f83c64e4ba31d03fc0ca97939b344013089211d7ff71e7" exitCode=0 Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.792068 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.928396 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume\") pod \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.928599 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume\") pod \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.928657 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4khzz\" (UniqueName: \"kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz\") pod \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\" (UID: \"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852\") " Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.929895 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume" (OuterVolumeSpecName: "config-volume") pod "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" (UID: "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.939035 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" (UID: "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 10:45:03 crc kubenswrapper[4733]: I0318 10:45:03.939176 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz" (OuterVolumeSpecName: "kube-api-access-4khzz") pod "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" (UID: "b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852"). InnerVolumeSpecName "kube-api-access-4khzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.030521 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4khzz\" (UniqueName: \"kubernetes.io/projected/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-kube-api-access-4khzz\") on node \"crc\" DevicePath \"\"" Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.030825 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.030956 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.450884 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" event={"ID":"b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852","Type":"ContainerDied","Data":"7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f"} Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.450934 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d997435b256176d51d88b14b3339d5330b41b331096ed74154ace289a620b7f" Mar 18 10:45:04 crc kubenswrapper[4733]: I0318 10:45:04.451408 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563845-8g7fj" Mar 18 10:45:06 crc kubenswrapper[4733]: I0318 10:45:06.175499 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:45:06 crc kubenswrapper[4733]: E0318 10:45:06.176549 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:45:13 crc kubenswrapper[4733]: I0318 10:45:13.179603 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:13 crc kubenswrapper[4733]: E0318 10:45:13.180960 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:18 crc kubenswrapper[4733]: I0318 10:45:18.175777 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:45:18 crc kubenswrapper[4733]: E0318 10:45:18.176397 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:45:24 crc kubenswrapper[4733]: I0318 10:45:24.176377 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:24 crc kubenswrapper[4733]: E0318 10:45:24.177960 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:32 crc kubenswrapper[4733]: I0318 10:45:32.177435 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:45:32 crc kubenswrapper[4733]: E0318 10:45:32.178511 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:45:35 crc kubenswrapper[4733]: I0318 10:45:35.175785 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:35 crc kubenswrapper[4733]: E0318 10:45:35.176789 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:46 crc kubenswrapper[4733]: I0318 10:45:46.175643 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:46 crc kubenswrapper[4733]: I0318 10:45:46.176394 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:45:46 crc kubenswrapper[4733]: E0318 10:45:46.176688 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:46 crc kubenswrapper[4733]: E0318 10:45:46.176810 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:45:57 crc kubenswrapper[4733]: I0318 10:45:57.175681 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:45:57 crc kubenswrapper[4733]: E0318 10:45:57.176879 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:45:58 crc kubenswrapper[4733]: I0318 10:45:58.175806 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:45:58 crc kubenswrapper[4733]: E0318 10:45:58.176384 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.164715 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563846-zfzc4"] Mar 18 10:46:00 crc kubenswrapper[4733]: E0318 10:46:00.165756 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" containerName="collect-profiles" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.165779 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" containerName="collect-profiles" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.166135 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43c9f9c-7dd6-45fd-b0e2-66e4e6de9852" containerName="collect-profiles" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.167059 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.171084 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.172843 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.174002 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.179410 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563846-zfzc4"] Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.187938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qcn\" (UniqueName: \"kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn\") pod \"auto-csr-approver-29563846-zfzc4\" (UID: \"02d9650e-918c-4ff4-82bd-ba01e08b6588\") " pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.289338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qcn\" (UniqueName: \"kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn\") pod \"auto-csr-approver-29563846-zfzc4\" (UID: \"02d9650e-918c-4ff4-82bd-ba01e08b6588\") " pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.323039 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qcn\" (UniqueName: \"kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn\") pod \"auto-csr-approver-29563846-zfzc4\" (UID: \"02d9650e-918c-4ff4-82bd-ba01e08b6588\") " pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.507619 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:00 crc kubenswrapper[4733]: I0318 10:46:00.959812 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563846-zfzc4"] Mar 18 10:46:00 crc kubenswrapper[4733]: W0318 10:46:00.966378 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02d9650e_918c_4ff4_82bd_ba01e08b6588.slice/crio-072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1 WatchSource:0}: Error finding container 072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1: Status 404 returned error can't find the container with id 072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1 Mar 18 10:46:01 crc kubenswrapper[4733]: I0318 10:46:01.084330 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" event={"ID":"02d9650e-918c-4ff4-82bd-ba01e08b6588","Type":"ContainerStarted","Data":"072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1"} Mar 18 10:46:04 crc kubenswrapper[4733]: I0318 10:46:04.118889 4733 generic.go:334] "Generic (PLEG): container finished" podID="02d9650e-918c-4ff4-82bd-ba01e08b6588" containerID="df6a0cb730f6fbfc1289efd16f5838442dcf854d748f1b313f239fe3a8ed31b9" exitCode=0 Mar 18 10:46:04 crc kubenswrapper[4733]: I0318 10:46:04.119036 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" event={"ID":"02d9650e-918c-4ff4-82bd-ba01e08b6588","Type":"ContainerDied","Data":"df6a0cb730f6fbfc1289efd16f5838442dcf854d748f1b313f239fe3a8ed31b9"} Mar 18 10:46:05 crc kubenswrapper[4733]: I0318 10:46:05.509704 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:05 crc kubenswrapper[4733]: I0318 10:46:05.681478 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2qcn\" (UniqueName: \"kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn\") pod \"02d9650e-918c-4ff4-82bd-ba01e08b6588\" (UID: \"02d9650e-918c-4ff4-82bd-ba01e08b6588\") " Mar 18 10:46:05 crc kubenswrapper[4733]: I0318 10:46:05.687402 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn" (OuterVolumeSpecName: "kube-api-access-s2qcn") pod "02d9650e-918c-4ff4-82bd-ba01e08b6588" (UID: "02d9650e-918c-4ff4-82bd-ba01e08b6588"). InnerVolumeSpecName "kube-api-access-s2qcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:46:05 crc kubenswrapper[4733]: I0318 10:46:05.782811 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2qcn\" (UniqueName: \"kubernetes.io/projected/02d9650e-918c-4ff4-82bd-ba01e08b6588-kube-api-access-s2qcn\") on node \"crc\" DevicePath \"\"" Mar 18 10:46:06 crc kubenswrapper[4733]: I0318 10:46:06.141452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" event={"ID":"02d9650e-918c-4ff4-82bd-ba01e08b6588","Type":"ContainerDied","Data":"072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1"} Mar 18 10:46:06 crc kubenswrapper[4733]: I0318 10:46:06.141498 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="072908d45ce872d36b19bfa216060105bf64d028952b77fda23a47fde15be9a1" Mar 18 10:46:06 crc kubenswrapper[4733]: I0318 10:46:06.141574 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563846-zfzc4" Mar 18 10:46:06 crc kubenswrapper[4733]: I0318 10:46:06.581996 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563840-lvjwz"] Mar 18 10:46:06 crc kubenswrapper[4733]: I0318 10:46:06.587717 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563840-lvjwz"] Mar 18 10:46:07 crc kubenswrapper[4733]: I0318 10:46:07.193857 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be59cc18-c769-443f-962f-042b8ba456b8" path="/var/lib/kubelet/pods/be59cc18-c769-443f-962f-042b8ba456b8/volumes" Mar 18 10:46:10 crc kubenswrapper[4733]: I0318 10:46:10.176379 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:46:10 crc kubenswrapper[4733]: E0318 10:46:10.178847 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:46:10 crc kubenswrapper[4733]: I0318 10:46:10.455459 4733 scope.go:117] "RemoveContainer" containerID="875056c9052fb9c1578503b6e7b5412bb4aa77c6a765ef7ede001dc3e0cb6698" Mar 18 10:46:11 crc kubenswrapper[4733]: I0318 10:46:11.210654 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:46:11 crc kubenswrapper[4733]: E0318 10:46:11.211093 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:46:24 crc kubenswrapper[4733]: I0318 10:46:24.175343 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:46:24 crc kubenswrapper[4733]: E0318 10:46:24.176527 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:46:25 crc kubenswrapper[4733]: I0318 10:46:25.176391 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:46:25 crc kubenswrapper[4733]: E0318 10:46:25.176952 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:46:37 crc kubenswrapper[4733]: I0318 10:46:37.175623 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:46:37 crc kubenswrapper[4733]: E0318 10:46:37.176671 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:46:40 crc kubenswrapper[4733]: I0318 10:46:40.176141 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:46:40 crc kubenswrapper[4733]: E0318 10:46:40.176979 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:46:48 crc kubenswrapper[4733]: I0318 10:46:48.176069 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:46:48 crc kubenswrapper[4733]: E0318 10:46:48.176946 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:46:54 crc kubenswrapper[4733]: I0318 10:46:54.176860 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:46:54 crc kubenswrapper[4733]: E0318 10:46:54.178034 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:47:02 crc kubenswrapper[4733]: I0318 10:47:02.176237 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:47:02 crc kubenswrapper[4733]: E0318 10:47:02.177529 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:47:06 crc kubenswrapper[4733]: I0318 10:47:06.175955 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:47:06 crc kubenswrapper[4733]: E0318 10:47:06.176686 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:47:13 crc kubenswrapper[4733]: I0318 10:47:13.570911 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:47:13 crc kubenswrapper[4733]: I0318 10:47:13.571704 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:47:14 crc kubenswrapper[4733]: I0318 10:47:14.176001 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:47:14 crc kubenswrapper[4733]: E0318 10:47:14.176416 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:47:21 crc kubenswrapper[4733]: I0318 10:47:21.184029 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:47:21 crc kubenswrapper[4733]: E0318 10:47:21.187343 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:47:26 crc kubenswrapper[4733]: I0318 10:47:26.175824 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:47:26 crc kubenswrapper[4733]: E0318 10:47:26.177110 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:47:34 crc kubenswrapper[4733]: I0318 10:47:34.175997 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:47:34 crc kubenswrapper[4733]: E0318 10:47:34.177043 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:47:40 crc kubenswrapper[4733]: I0318 10:47:40.176579 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:47:40 crc kubenswrapper[4733]: E0318 10:47:40.177800 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:47:43 crc kubenswrapper[4733]: I0318 10:47:43.571738 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:47:43 crc kubenswrapper[4733]: I0318 10:47:43.572181 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.711144 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:47:46 crc kubenswrapper[4733]: E0318 10:47:46.712126 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d9650e-918c-4ff4-82bd-ba01e08b6588" containerName="oc" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.712149 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d9650e-918c-4ff4-82bd-ba01e08b6588" containerName="oc" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.712513 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d9650e-918c-4ff4-82bd-ba01e08b6588" containerName="oc" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.714527 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.719176 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.770439 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.770665 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.770748 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfbkb\" (UniqueName: \"kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.872331 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.872396 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfbkb\" (UniqueName: \"kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.872424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.873054 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.873081 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:46 crc kubenswrapper[4733]: I0318 10:47:46.903286 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfbkb\" (UniqueName: \"kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb\") pod \"community-operators-7fmwk\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:47 crc kubenswrapper[4733]: I0318 10:47:47.040576 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:47 crc kubenswrapper[4733]: I0318 10:47:47.358856 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:47:48 crc kubenswrapper[4733]: I0318 10:47:48.175710 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:47:48 crc kubenswrapper[4733]: E0318 10:47:48.176966 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:47:48 crc kubenswrapper[4733]: I0318 10:47:48.258450 4733 generic.go:334] "Generic (PLEG): container finished" podID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerID="7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27" exitCode=0 Mar 18 10:47:48 crc kubenswrapper[4733]: I0318 10:47:48.258533 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerDied","Data":"7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27"} Mar 18 10:47:48 crc kubenswrapper[4733]: I0318 10:47:48.258620 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerStarted","Data":"928d80f6e5ee50442d6ebfaa23c4a5899ef5981249f762df7a2c9401fd78c54d"} Mar 18 10:47:49 crc kubenswrapper[4733]: I0318 10:47:49.267706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerStarted","Data":"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee"} Mar 18 10:47:50 crc kubenswrapper[4733]: I0318 10:47:50.281098 4733 generic.go:334] "Generic (PLEG): container finished" podID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerID="c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee" exitCode=0 Mar 18 10:47:50 crc kubenswrapper[4733]: I0318 10:47:50.281176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerDied","Data":"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee"} Mar 18 10:47:51 crc kubenswrapper[4733]: I0318 10:47:51.315667 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerStarted","Data":"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d"} Mar 18 10:47:51 crc kubenswrapper[4733]: I0318 10:47:51.345280 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7fmwk" podStartSLOduration=2.912951338 podStartE2EDuration="5.34525992s" podCreationTimestamp="2026-03-18 10:47:46 +0000 UTC" firstStartedPulling="2026-03-18 10:47:48.261231371 +0000 UTC m=+2107.752965726" lastFinishedPulling="2026-03-18 10:47:50.693539953 +0000 UTC m=+2110.185274308" observedRunningTime="2026-03-18 10:47:51.341830333 +0000 UTC m=+2110.833564698" watchObservedRunningTime="2026-03-18 10:47:51.34525992 +0000 UTC m=+2110.836994255" Mar 18 10:47:54 crc kubenswrapper[4733]: I0318 10:47:54.175514 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:47:54 crc kubenswrapper[4733]: E0318 10:47:54.176261 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:47:57 crc kubenswrapper[4733]: I0318 10:47:57.041609 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:57 crc kubenswrapper[4733]: I0318 10:47:57.042016 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:57 crc kubenswrapper[4733]: I0318 10:47:57.121141 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:57 crc kubenswrapper[4733]: I0318 10:47:57.441375 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:47:57 crc kubenswrapper[4733]: I0318 10:47:57.498627 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:47:59 crc kubenswrapper[4733]: I0318 10:47:59.385859 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7fmwk" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="registry-server" containerID="cri-o://19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d" gracePeriod=2 Mar 18 10:47:59 crc kubenswrapper[4733]: I0318 10:47:59.910825 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.111363 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities\") pod \"73638d51-1c01-4323-91f7-ef8fba5f8654\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.111503 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfbkb\" (UniqueName: \"kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb\") pod \"73638d51-1c01-4323-91f7-ef8fba5f8654\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.111660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content\") pod \"73638d51-1c01-4323-91f7-ef8fba5f8654\" (UID: \"73638d51-1c01-4323-91f7-ef8fba5f8654\") " Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.112595 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities" (OuterVolumeSpecName: "utilities") pod "73638d51-1c01-4323-91f7-ef8fba5f8654" (UID: "73638d51-1c01-4323-91f7-ef8fba5f8654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.131857 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb" (OuterVolumeSpecName: "kube-api-access-vfbkb") pod "73638d51-1c01-4323-91f7-ef8fba5f8654" (UID: "73638d51-1c01-4323-91f7-ef8fba5f8654"). InnerVolumeSpecName "kube-api-access-vfbkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.163323 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563848-2tfjq"] Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.167173 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="registry-server" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.167380 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="registry-server" Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.167492 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="extract-utilities" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.167591 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="extract-utilities" Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.167698 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="extract-content" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.167799 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="extract-content" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.169583 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerName="registry-server" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.170627 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.173940 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.174347 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.174564 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.181385 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563848-2tfjq"] Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.190989 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73638d51-1c01-4323-91f7-ef8fba5f8654" (UID: "73638d51-1c01-4323-91f7-ef8fba5f8654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.214154 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.214219 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfbkb\" (UniqueName: \"kubernetes.io/projected/73638d51-1c01-4323-91f7-ef8fba5f8654-kube-api-access-vfbkb\") on node \"crc\" DevicePath \"\"" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.214234 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73638d51-1c01-4323-91f7-ef8fba5f8654-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.315544 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x77ls\" (UniqueName: \"kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls\") pod \"auto-csr-approver-29563848-2tfjq\" (UID: \"1bc1a1de-2965-422f-8ac4-77358d0d7df6\") " pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.395975 4733 generic.go:334] "Generic (PLEG): container finished" podID="73638d51-1c01-4323-91f7-ef8fba5f8654" containerID="19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d" exitCode=0 Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.396034 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerDied","Data":"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d"} Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.396057 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fmwk" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.396073 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fmwk" event={"ID":"73638d51-1c01-4323-91f7-ef8fba5f8654","Type":"ContainerDied","Data":"928d80f6e5ee50442d6ebfaa23c4a5899ef5981249f762df7a2c9401fd78c54d"} Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.396107 4733 scope.go:117] "RemoveContainer" containerID="19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.423328 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x77ls\" (UniqueName: \"kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls\") pod \"auto-csr-approver-29563848-2tfjq\" (UID: \"1bc1a1de-2965-422f-8ac4-77358d0d7df6\") " pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.439743 4733 scope.go:117] "RemoveContainer" containerID="c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.445838 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.458754 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7fmwk"] Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.459106 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x77ls\" (UniqueName: \"kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls\") pod \"auto-csr-approver-29563848-2tfjq\" (UID: \"1bc1a1de-2965-422f-8ac4-77358d0d7df6\") " pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.477962 4733 scope.go:117] "RemoveContainer" containerID="7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.519963 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.522286 4733 scope.go:117] "RemoveContainer" containerID="19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d" Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.522776 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d\": container with ID starting with 19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d not found: ID does not exist" containerID="19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.522813 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d"} err="failed to get container status \"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d\": rpc error: code = NotFound desc = could not find container \"19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d\": container with ID starting with 19e7075c2d39be3d5800b88e9f5f6ed50e174eda4a0526fd16da79a0781fe40d not found: ID does not exist" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.522863 4733 scope.go:117] "RemoveContainer" containerID="c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee" Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.523177 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee\": container with ID starting with c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee not found: ID does not exist" containerID="c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.523221 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee"} err="failed to get container status \"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee\": rpc error: code = NotFound desc = could not find container \"c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee\": container with ID starting with c5107691d43bccc1194ae8e4cd1cf5beed33bacfad0b1a00566988d322aecaee not found: ID does not exist" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.523245 4733 scope.go:117] "RemoveContainer" containerID="7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27" Mar 18 10:48:00 crc kubenswrapper[4733]: E0318 10:48:00.523591 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27\": container with ID starting with 7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27 not found: ID does not exist" containerID="7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27" Mar 18 10:48:00 crc kubenswrapper[4733]: I0318 10:48:00.523624 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27"} err="failed to get container status \"7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27\": rpc error: code = NotFound desc = could not find container \"7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27\": container with ID starting with 7c559db9726092272a44b539af5e62593a410a2361a733d6503432107bd1fe27 not found: ID does not exist" Mar 18 10:48:01 crc kubenswrapper[4733]: W0318 10:48:01.042962 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bc1a1de_2965_422f_8ac4_77358d0d7df6.slice/crio-23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6 WatchSource:0}: Error finding container 23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6: Status 404 returned error can't find the container with id 23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6 Mar 18 10:48:01 crc kubenswrapper[4733]: I0318 10:48:01.047821 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563848-2tfjq"] Mar 18 10:48:01 crc kubenswrapper[4733]: I0318 10:48:01.184661 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:48:01 crc kubenswrapper[4733]: E0318 10:48:01.184993 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:48:01 crc kubenswrapper[4733]: I0318 10:48:01.188988 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73638d51-1c01-4323-91f7-ef8fba5f8654" path="/var/lib/kubelet/pods/73638d51-1c01-4323-91f7-ef8fba5f8654/volumes" Mar 18 10:48:01 crc kubenswrapper[4733]: I0318 10:48:01.411475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" event={"ID":"1bc1a1de-2965-422f-8ac4-77358d0d7df6","Type":"ContainerStarted","Data":"23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6"} Mar 18 10:48:02 crc kubenswrapper[4733]: I0318 10:48:02.420972 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" event={"ID":"1bc1a1de-2965-422f-8ac4-77358d0d7df6","Type":"ContainerStarted","Data":"6fe3d8a40e1ba17924153a168a35cf8fd5e9cc3d1fdeb9b0f70b81b8350f5f56"} Mar 18 10:48:02 crc kubenswrapper[4733]: I0318 10:48:02.435980 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" podStartSLOduration=1.613556582 podStartE2EDuration="2.435958701s" podCreationTimestamp="2026-03-18 10:48:00 +0000 UTC" firstStartedPulling="2026-03-18 10:48:01.048358665 +0000 UTC m=+2120.540093030" lastFinishedPulling="2026-03-18 10:48:01.870760794 +0000 UTC m=+2121.362495149" observedRunningTime="2026-03-18 10:48:02.434074088 +0000 UTC m=+2121.925808423" watchObservedRunningTime="2026-03-18 10:48:02.435958701 +0000 UTC m=+2121.927693036" Mar 18 10:48:03 crc kubenswrapper[4733]: I0318 10:48:03.433392 4733 generic.go:334] "Generic (PLEG): container finished" podID="1bc1a1de-2965-422f-8ac4-77358d0d7df6" containerID="6fe3d8a40e1ba17924153a168a35cf8fd5e9cc3d1fdeb9b0f70b81b8350f5f56" exitCode=0 Mar 18 10:48:03 crc kubenswrapper[4733]: I0318 10:48:03.433438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" event={"ID":"1bc1a1de-2965-422f-8ac4-77358d0d7df6","Type":"ContainerDied","Data":"6fe3d8a40e1ba17924153a168a35cf8fd5e9cc3d1fdeb9b0f70b81b8350f5f56"} Mar 18 10:48:04 crc kubenswrapper[4733]: I0318 10:48:04.780760 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:04 crc kubenswrapper[4733]: I0318 10:48:04.901017 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x77ls\" (UniqueName: \"kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls\") pod \"1bc1a1de-2965-422f-8ac4-77358d0d7df6\" (UID: \"1bc1a1de-2965-422f-8ac4-77358d0d7df6\") " Mar 18 10:48:04 crc kubenswrapper[4733]: I0318 10:48:04.909972 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls" (OuterVolumeSpecName: "kube-api-access-x77ls") pod "1bc1a1de-2965-422f-8ac4-77358d0d7df6" (UID: "1bc1a1de-2965-422f-8ac4-77358d0d7df6"). InnerVolumeSpecName "kube-api-access-x77ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.003019 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x77ls\" (UniqueName: \"kubernetes.io/projected/1bc1a1de-2965-422f-8ac4-77358d0d7df6-kube-api-access-x77ls\") on node \"crc\" DevicePath \"\"" Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.453274 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" event={"ID":"1bc1a1de-2965-422f-8ac4-77358d0d7df6","Type":"ContainerDied","Data":"23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6"} Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.453335 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23a710f8d256ee258f90aa54f124b4ae12454a3723d975a0640cc7e9a315b4e6" Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.453355 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563848-2tfjq" Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.541273 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563842-78hdh"] Mar 18 10:48:05 crc kubenswrapper[4733]: I0318 10:48:05.552639 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563842-78hdh"] Mar 18 10:48:06 crc kubenswrapper[4733]: I0318 10:48:06.176681 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:48:06 crc kubenswrapper[4733]: E0318 10:48:06.177071 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:48:07 crc kubenswrapper[4733]: I0318 10:48:07.191578 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6751ff-c17a-472f-b315-06edff233f07" path="/var/lib/kubelet/pods/8a6751ff-c17a-472f-b315-06edff233f07/volumes" Mar 18 10:48:10 crc kubenswrapper[4733]: I0318 10:48:10.583357 4733 scope.go:117] "RemoveContainer" containerID="56204ccc444d383e182644d7909ef75775c83c4da5e6940b6afcaf6c25fa0fc2" Mar 18 10:48:13 crc kubenswrapper[4733]: I0318 10:48:13.575914 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:48:13 crc kubenswrapper[4733]: I0318 10:48:13.576607 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:48:13 crc kubenswrapper[4733]: I0318 10:48:13.576675 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:48:13 crc kubenswrapper[4733]: I0318 10:48:13.577588 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:48:13 crc kubenswrapper[4733]: I0318 10:48:13.577687 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337" gracePeriod=600 Mar 18 10:48:14 crc kubenswrapper[4733]: I0318 10:48:14.175999 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:48:14 crc kubenswrapper[4733]: E0318 10:48:14.177114 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:48:14 crc kubenswrapper[4733]: I0318 10:48:14.536706 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337" exitCode=0 Mar 18 10:48:14 crc kubenswrapper[4733]: I0318 10:48:14.536747 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337"} Mar 18 10:48:14 crc kubenswrapper[4733]: I0318 10:48:14.536773 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48"} Mar 18 10:48:14 crc kubenswrapper[4733]: I0318 10:48:14.536789 4733 scope.go:117] "RemoveContainer" containerID="fc33062a38e6003bcfe678b0b641bcd73299a07f8dcc32e6f590e8bb7c29b637" Mar 18 10:48:21 crc kubenswrapper[4733]: I0318 10:48:21.179792 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:48:21 crc kubenswrapper[4733]: E0318 10:48:21.180560 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:48:26 crc kubenswrapper[4733]: I0318 10:48:26.176926 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:48:26 crc kubenswrapper[4733]: I0318 10:48:26.671917 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60"} Mar 18 10:48:26 crc kubenswrapper[4733]: I0318 10:48:26.673120 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:48:30 crc kubenswrapper[4733]: I0318 10:48:30.709972 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" exitCode=0 Mar 18 10:48:30 crc kubenswrapper[4733]: I0318 10:48:30.710102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60"} Mar 18 10:48:30 crc kubenswrapper[4733]: I0318 10:48:30.710530 4733 scope.go:117] "RemoveContainer" containerID="c6feb5dbe0273076fcc0fc6ebcba5d0f6774752ba12f9bdf9dbed3f20d646a98" Mar 18 10:48:30 crc kubenswrapper[4733]: I0318 10:48:30.711421 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:48:30 crc kubenswrapper[4733]: E0318 10:48:30.711792 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:48:32 crc kubenswrapper[4733]: I0318 10:48:32.176218 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:48:32 crc kubenswrapper[4733]: I0318 10:48:32.735312 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f"} Mar 18 10:48:32 crc kubenswrapper[4733]: I0318 10:48:32.735832 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:48:37 crc kubenswrapper[4733]: I0318 10:48:37.779537 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" exitCode=0 Mar 18 10:48:37 crc kubenswrapper[4733]: I0318 10:48:37.779620 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f"} Mar 18 10:48:37 crc kubenswrapper[4733]: I0318 10:48:37.779842 4733 scope.go:117] "RemoveContainer" containerID="5270a9bb2caacb712ce1222dbbad48bb5b8a42db385b8791217d8ea1eab704e8" Mar 18 10:48:37 crc kubenswrapper[4733]: I0318 10:48:37.780658 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:48:37 crc kubenswrapper[4733]: E0318 10:48:37.781252 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:48:43 crc kubenswrapper[4733]: I0318 10:48:43.176383 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:48:43 crc kubenswrapper[4733]: E0318 10:48:43.177363 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:48:51 crc kubenswrapper[4733]: I0318 10:48:51.187069 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:48:51 crc kubenswrapper[4733]: E0318 10:48:51.188426 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:48:57 crc kubenswrapper[4733]: I0318 10:48:57.176583 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:48:57 crc kubenswrapper[4733]: E0318 10:48:57.177854 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:49:02 crc kubenswrapper[4733]: I0318 10:49:02.176826 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:49:02 crc kubenswrapper[4733]: E0318 10:49:02.177510 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:49:10 crc kubenswrapper[4733]: I0318 10:49:10.176168 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:49:10 crc kubenswrapper[4733]: E0318 10:49:10.177301 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:49:14 crc kubenswrapper[4733]: I0318 10:49:14.176704 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:49:14 crc kubenswrapper[4733]: E0318 10:49:14.177778 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:49:23 crc kubenswrapper[4733]: I0318 10:49:23.176499 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:49:23 crc kubenswrapper[4733]: E0318 10:49:23.177831 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:49:28 crc kubenswrapper[4733]: I0318 10:49:28.175665 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:49:28 crc kubenswrapper[4733]: E0318 10:49:28.176881 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:49:38 crc kubenswrapper[4733]: I0318 10:49:38.175699 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:49:38 crc kubenswrapper[4733]: E0318 10:49:38.176675 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:49:40 crc kubenswrapper[4733]: I0318 10:49:40.175928 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:49:40 crc kubenswrapper[4733]: E0318 10:49:40.176991 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.634946 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:49:47 crc kubenswrapper[4733]: E0318 10:49:47.636483 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1a1de-2965-422f-8ac4-77358d0d7df6" containerName="oc" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.636506 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1a1de-2965-422f-8ac4-77358d0d7df6" containerName="oc" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.637019 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1a1de-2965-422f-8ac4-77358d0d7df6" containerName="oc" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.641996 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.692314 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.755852 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h8vm\" (UniqueName: \"kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.755960 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.755993 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.857975 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h8vm\" (UniqueName: \"kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.858084 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.858122 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.858723 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.858869 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:47 crc kubenswrapper[4733]: I0318 10:49:47.880816 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h8vm\" (UniqueName: \"kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm\") pod \"redhat-operators-2nkxk\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.002512 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.220543 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.223502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.240049 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.365227 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.365568 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxdl\" (UniqueName: \"kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.365702 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.469353 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.470302 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxdl\" (UniqueName: \"kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.470394 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.470416 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.470982 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.471385 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.494065 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxdl\" (UniqueName: \"kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl\") pod \"redhat-marketplace-p567z\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.500605 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerStarted","Data":"ecd120dd90e1036bb5e6d554a5806a7bd302457439032a59781bedfcd41b555a"} Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.540502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:48 crc kubenswrapper[4733]: I0318 10:49:48.960584 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.509103 4733 generic.go:334] "Generic (PLEG): container finished" podID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerID="716b0e092d29f01bae19a735d99fb8b66ebdd4dfe0eb107d62d5d4661ad290b4" exitCode=0 Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.509180 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerDied","Data":"716b0e092d29f01bae19a735d99fb8b66ebdd4dfe0eb107d62d5d4661ad290b4"} Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.509271 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerStarted","Data":"5941ae8ebdec11816d7bbdab01c89fa7a8cb1647d4a432cd5fc43d5b8d08123c"} Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.511053 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerID="c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42" exitCode=0 Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.511079 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerDied","Data":"c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42"} Mar 18 10:49:49 crc kubenswrapper[4733]: I0318 10:49:49.511565 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:49:50 crc kubenswrapper[4733]: I0318 10:49:50.176127 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:49:50 crc kubenswrapper[4733]: E0318 10:49:50.176946 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:49:50 crc kubenswrapper[4733]: I0318 10:49:50.519812 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerStarted","Data":"d6b6c73dfb062b43a2fff48a697f356e2654828365f2c4918111e23d64e38e06"} Mar 18 10:49:50 crc kubenswrapper[4733]: I0318 10:49:50.525266 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerStarted","Data":"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7"} Mar 18 10:49:51 crc kubenswrapper[4733]: I0318 10:49:51.182280 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:49:51 crc kubenswrapper[4733]: E0318 10:49:51.183267 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:49:51 crc kubenswrapper[4733]: I0318 10:49:51.538288 4733 generic.go:334] "Generic (PLEG): container finished" podID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerID="d6b6c73dfb062b43a2fff48a697f356e2654828365f2c4918111e23d64e38e06" exitCode=0 Mar 18 10:49:51 crc kubenswrapper[4733]: I0318 10:49:51.538369 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerDied","Data":"d6b6c73dfb062b43a2fff48a697f356e2654828365f2c4918111e23d64e38e06"} Mar 18 10:49:51 crc kubenswrapper[4733]: I0318 10:49:51.540974 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerID="c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7" exitCode=0 Mar 18 10:49:51 crc kubenswrapper[4733]: I0318 10:49:51.541030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerDied","Data":"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7"} Mar 18 10:49:52 crc kubenswrapper[4733]: I0318 10:49:52.553830 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerStarted","Data":"a2de23c37da85617652573e8df45d4ddd82910052cd6fefb446ea209a4de9185"} Mar 18 10:49:52 crc kubenswrapper[4733]: I0318 10:49:52.557334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerStarted","Data":"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262"} Mar 18 10:49:52 crc kubenswrapper[4733]: I0318 10:49:52.595041 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p567z" podStartSLOduration=2.018794697 podStartE2EDuration="4.595012614s" podCreationTimestamp="2026-03-18 10:49:48 +0000 UTC" firstStartedPulling="2026-03-18 10:49:49.511149056 +0000 UTC m=+2229.002883421" lastFinishedPulling="2026-03-18 10:49:52.087366973 +0000 UTC m=+2231.579101338" observedRunningTime="2026-03-18 10:49:52.581171234 +0000 UTC m=+2232.072905559" watchObservedRunningTime="2026-03-18 10:49:52.595012614 +0000 UTC m=+2232.086746979" Mar 18 10:49:52 crc kubenswrapper[4733]: I0318 10:49:52.604076 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2nkxk" podStartSLOduration=3.1111192 podStartE2EDuration="5.604056099s" podCreationTimestamp="2026-03-18 10:49:47 +0000 UTC" firstStartedPulling="2026-03-18 10:49:49.513836512 +0000 UTC m=+2229.005570877" lastFinishedPulling="2026-03-18 10:49:52.006773421 +0000 UTC m=+2231.498507776" observedRunningTime="2026-03-18 10:49:52.597989848 +0000 UTC m=+2232.089724193" watchObservedRunningTime="2026-03-18 10:49:52.604056099 +0000 UTC m=+2232.095790434" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.003059 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.003697 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.541154 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.541920 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.599097 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.680937 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:49:58 crc kubenswrapper[4733]: I0318 10:49:58.842463 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:49:59 crc kubenswrapper[4733]: I0318 10:49:59.062163 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2nkxk" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="registry-server" probeResult="failure" output=< Mar 18 10:49:59 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 10:49:59 crc kubenswrapper[4733]: > Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.153644 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563850-8gwpw"] Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.155311 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.159252 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.159818 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.160769 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.165671 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563850-8gwpw"] Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.266847 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ww9k\" (UniqueName: \"kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k\") pod \"auto-csr-approver-29563850-8gwpw\" (UID: \"2166cf23-2a65-4b17-922e-3131be1d6d8b\") " pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.369466 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ww9k\" (UniqueName: \"kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k\") pod \"auto-csr-approver-29563850-8gwpw\" (UID: \"2166cf23-2a65-4b17-922e-3131be1d6d8b\") " pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.397695 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ww9k\" (UniqueName: \"kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k\") pod \"auto-csr-approver-29563850-8gwpw\" (UID: \"2166cf23-2a65-4b17-922e-3131be1d6d8b\") " pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.497075 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:00 crc kubenswrapper[4733]: I0318 10:50:00.678682 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p567z" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="registry-server" containerID="cri-o://a2de23c37da85617652573e8df45d4ddd82910052cd6fefb446ea209a4de9185" gracePeriod=2 Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.042318 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563850-8gwpw"] Mar 18 10:50:01 crc kubenswrapper[4733]: W0318 10:50:01.053144 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2166cf23_2a65_4b17_922e_3131be1d6d8b.slice/crio-46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec WatchSource:0}: Error finding container 46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec: Status 404 returned error can't find the container with id 46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.185782 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:50:01 crc kubenswrapper[4733]: E0318 10:50:01.186411 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.690454 4733 generic.go:334] "Generic (PLEG): container finished" podID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerID="a2de23c37da85617652573e8df45d4ddd82910052cd6fefb446ea209a4de9185" exitCode=0 Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.690524 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerDied","Data":"a2de23c37da85617652573e8df45d4ddd82910052cd6fefb446ea209a4de9185"} Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.693406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" event={"ID":"2166cf23-2a65-4b17-922e-3131be1d6d8b","Type":"ContainerStarted","Data":"46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec"} Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.817473 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.916717 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gxdl\" (UniqueName: \"kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl\") pod \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.916842 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities\") pod \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.917028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content\") pod \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\" (UID: \"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896\") " Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.917776 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities" (OuterVolumeSpecName: "utilities") pod "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" (UID: "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.926386 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl" (OuterVolumeSpecName: "kube-api-access-6gxdl") pod "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" (UID: "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896"). InnerVolumeSpecName "kube-api-access-6gxdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:50:01 crc kubenswrapper[4733]: I0318 10:50:01.940309 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" (UID: "b49b68f2-ffd0-4ff7-a4ae-691c0a71e896"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.018415 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.018444 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gxdl\" (UniqueName: \"kubernetes.io/projected/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-kube-api-access-6gxdl\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.018455 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.175861 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:50:02 crc kubenswrapper[4733]: E0318 10:50:02.176261 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.701818 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p567z" event={"ID":"b49b68f2-ffd0-4ff7-a4ae-691c0a71e896","Type":"ContainerDied","Data":"5941ae8ebdec11816d7bbdab01c89fa7a8cb1647d4a432cd5fc43d5b8d08123c"} Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.702113 4733 scope.go:117] "RemoveContainer" containerID="a2de23c37da85617652573e8df45d4ddd82910052cd6fefb446ea209a4de9185" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.701851 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p567z" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.704548 4733 generic.go:334] "Generic (PLEG): container finished" podID="2166cf23-2a65-4b17-922e-3131be1d6d8b" containerID="649f36c8155821a228e5fee55c54c1d5edbde655cee7563ac249384dedef675b" exitCode=0 Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.704606 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" event={"ID":"2166cf23-2a65-4b17-922e-3131be1d6d8b","Type":"ContainerDied","Data":"649f36c8155821a228e5fee55c54c1d5edbde655cee7563ac249384dedef675b"} Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.731919 4733 scope.go:117] "RemoveContainer" containerID="d6b6c73dfb062b43a2fff48a697f356e2654828365f2c4918111e23d64e38e06" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.761279 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.763221 4733 scope.go:117] "RemoveContainer" containerID="716b0e092d29f01bae19a735d99fb8b66ebdd4dfe0eb107d62d5d4661ad290b4" Mar 18 10:50:02 crc kubenswrapper[4733]: I0318 10:50:02.768687 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p567z"] Mar 18 10:50:03 crc kubenswrapper[4733]: I0318 10:50:03.185246 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" path="/var/lib/kubelet/pods/b49b68f2-ffd0-4ff7-a4ae-691c0a71e896/volumes" Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.113682 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.266839 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ww9k\" (UniqueName: \"kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k\") pod \"2166cf23-2a65-4b17-922e-3131be1d6d8b\" (UID: \"2166cf23-2a65-4b17-922e-3131be1d6d8b\") " Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.283099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k" (OuterVolumeSpecName: "kube-api-access-2ww9k") pod "2166cf23-2a65-4b17-922e-3131be1d6d8b" (UID: "2166cf23-2a65-4b17-922e-3131be1d6d8b"). InnerVolumeSpecName "kube-api-access-2ww9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.371124 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ww9k\" (UniqueName: \"kubernetes.io/projected/2166cf23-2a65-4b17-922e-3131be1d6d8b-kube-api-access-2ww9k\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.734327 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" event={"ID":"2166cf23-2a65-4b17-922e-3131be1d6d8b","Type":"ContainerDied","Data":"46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec"} Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.734387 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46ecf4faa11e9d76da828b425e3870304d56ac67fef7dbfbe7f42a0695dc20ec" Mar 18 10:50:04 crc kubenswrapper[4733]: I0318 10:50:04.734399 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563850-8gwpw" Mar 18 10:50:05 crc kubenswrapper[4733]: I0318 10:50:05.219047 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563844-cdw7c"] Mar 18 10:50:05 crc kubenswrapper[4733]: I0318 10:50:05.230677 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563844-cdw7c"] Mar 18 10:50:07 crc kubenswrapper[4733]: I0318 10:50:07.193947 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a764647f-75c8-4ce3-82fb-b2b729a895a0" path="/var/lib/kubelet/pods/a764647f-75c8-4ce3-82fb-b2b729a895a0/volumes" Mar 18 10:50:08 crc kubenswrapper[4733]: I0318 10:50:08.087878 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:50:08 crc kubenswrapper[4733]: I0318 10:50:08.174350 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:50:08 crc kubenswrapper[4733]: I0318 10:50:08.339467 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:50:09 crc kubenswrapper[4733]: I0318 10:50:09.787849 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2nkxk" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="registry-server" containerID="cri-o://5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262" gracePeriod=2 Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.389984 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.502068 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h8vm\" (UniqueName: \"kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm\") pod \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.502218 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content\") pod \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.502385 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities\") pod \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\" (UID: \"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2\") " Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.504110 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities" (OuterVolumeSpecName: "utilities") pod "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" (UID: "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.508845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm" (OuterVolumeSpecName: "kube-api-access-5h8vm") pod "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" (UID: "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2"). InnerVolumeSpecName "kube-api-access-5h8vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.604476 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.604523 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h8vm\" (UniqueName: \"kubernetes.io/projected/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-kube-api-access-5h8vm\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.674340 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" (UID: "c9c6e055-5d64-4158-b0d2-27de0cb9f7c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.706855 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.720951 4733 scope.go:117] "RemoveContainer" containerID="4e31a574451b3bae3c66b0663fefb42d9a6b941c8dd7f0cf1a6c603f449c0e3b" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.799365 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerID="5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262" exitCode=0 Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.799416 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerDied","Data":"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262"} Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.799453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nkxk" event={"ID":"c9c6e055-5d64-4158-b0d2-27de0cb9f7c2","Type":"ContainerDied","Data":"ecd120dd90e1036bb5e6d554a5806a7bd302457439032a59781bedfcd41b555a"} Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.799458 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nkxk" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.799471 4733 scope.go:117] "RemoveContainer" containerID="5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.827059 4733 scope.go:117] "RemoveContainer" containerID="c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.843293 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.850765 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2nkxk"] Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.866848 4733 scope.go:117] "RemoveContainer" containerID="c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.889370 4733 scope.go:117] "RemoveContainer" containerID="5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262" Mar 18 10:50:10 crc kubenswrapper[4733]: E0318 10:50:10.889782 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262\": container with ID starting with 5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262 not found: ID does not exist" containerID="5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.889843 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262"} err="failed to get container status \"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262\": rpc error: code = NotFound desc = could not find container \"5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262\": container with ID starting with 5a8d6935c3c6f86eaf07e3b76ff0680d7842eea801abcd908667011593413262 not found: ID does not exist" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.889877 4733 scope.go:117] "RemoveContainer" containerID="c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7" Mar 18 10:50:10 crc kubenswrapper[4733]: E0318 10:50:10.890293 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7\": container with ID starting with c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7 not found: ID does not exist" containerID="c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.890336 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7"} err="failed to get container status \"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7\": rpc error: code = NotFound desc = could not find container \"c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7\": container with ID starting with c30c08fe95cea2bf662162f1cab4d1b78eb0e6f559bd1de70cda2437fd791ae7 not found: ID does not exist" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.890367 4733 scope.go:117] "RemoveContainer" containerID="c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42" Mar 18 10:50:10 crc kubenswrapper[4733]: E0318 10:50:10.890699 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42\": container with ID starting with c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42 not found: ID does not exist" containerID="c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42" Mar 18 10:50:10 crc kubenswrapper[4733]: I0318 10:50:10.890733 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42"} err="failed to get container status \"c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42\": rpc error: code = NotFound desc = could not find container \"c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42\": container with ID starting with c629fb05ebbbd1dadfdfa794e8f50fea898971c063ea18eb304ec5d519dceb42 not found: ID does not exist" Mar 18 10:50:11 crc kubenswrapper[4733]: I0318 10:50:11.191584 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" path="/var/lib/kubelet/pods/c9c6e055-5d64-4158-b0d2-27de0cb9f7c2/volumes" Mar 18 10:50:13 crc kubenswrapper[4733]: I0318 10:50:13.571138 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:50:13 crc kubenswrapper[4733]: I0318 10:50:13.571554 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:50:15 crc kubenswrapper[4733]: I0318 10:50:15.176481 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:50:15 crc kubenswrapper[4733]: E0318 10:50:15.177185 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:50:17 crc kubenswrapper[4733]: I0318 10:50:17.175754 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:50:17 crc kubenswrapper[4733]: E0318 10:50:17.176167 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:50:28 crc kubenswrapper[4733]: I0318 10:50:28.175406 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:50:28 crc kubenswrapper[4733]: I0318 10:50:28.176000 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:50:28 crc kubenswrapper[4733]: E0318 10:50:28.176262 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:50:28 crc kubenswrapper[4733]: E0318 10:50:28.176491 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:50:40 crc kubenswrapper[4733]: I0318 10:50:40.175827 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:50:40 crc kubenswrapper[4733]: E0318 10:50:40.177813 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:50:42 crc kubenswrapper[4733]: I0318 10:50:42.175101 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:50:42 crc kubenswrapper[4733]: E0318 10:50:42.175410 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:50:43 crc kubenswrapper[4733]: I0318 10:50:43.571034 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:50:43 crc kubenswrapper[4733]: I0318 10:50:43.571562 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:50:52 crc kubenswrapper[4733]: I0318 10:50:52.175966 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:50:52 crc kubenswrapper[4733]: E0318 10:50:52.177176 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:50:53 crc kubenswrapper[4733]: I0318 10:50:53.176730 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:50:53 crc kubenswrapper[4733]: E0318 10:50:53.177385 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:51:04 crc kubenswrapper[4733]: I0318 10:51:04.175969 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:51:04 crc kubenswrapper[4733]: I0318 10:51:04.176622 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:51:04 crc kubenswrapper[4733]: E0318 10:51:04.176820 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:51:04 crc kubenswrapper[4733]: E0318 10:51:04.177072 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:51:13 crc kubenswrapper[4733]: I0318 10:51:13.570817 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:51:13 crc kubenswrapper[4733]: I0318 10:51:13.571437 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:51:13 crc kubenswrapper[4733]: I0318 10:51:13.571488 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:51:13 crc kubenswrapper[4733]: I0318 10:51:13.572228 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:51:13 crc kubenswrapper[4733]: I0318 10:51:13.572304 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" gracePeriod=600 Mar 18 10:51:13 crc kubenswrapper[4733]: E0318 10:51:13.718094 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:51:14 crc kubenswrapper[4733]: I0318 10:51:14.196162 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" exitCode=0 Mar 18 10:51:14 crc kubenswrapper[4733]: I0318 10:51:14.196241 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48"} Mar 18 10:51:14 crc kubenswrapper[4733]: I0318 10:51:14.196596 4733 scope.go:117] "RemoveContainer" containerID="4aabe714853c502719880f7f27bb562465a6a84fdec9e321e389ec23753f6337" Mar 18 10:51:14 crc kubenswrapper[4733]: I0318 10:51:14.197263 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:51:14 crc kubenswrapper[4733]: E0318 10:51:14.198489 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:51:15 crc kubenswrapper[4733]: I0318 10:51:15.176925 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:51:15 crc kubenswrapper[4733]: E0318 10:51:15.177149 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:51:16 crc kubenswrapper[4733]: I0318 10:51:16.176574 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:51:16 crc kubenswrapper[4733]: E0318 10:51:16.177533 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:51:18 crc kubenswrapper[4733]: E0318 10:51:18.206961 4733 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.032s" Mar 18 10:51:26 crc kubenswrapper[4733]: I0318 10:51:26.176084 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:51:26 crc kubenswrapper[4733]: E0318 10:51:26.177029 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:51:28 crc kubenswrapper[4733]: I0318 10:51:28.175958 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:51:28 crc kubenswrapper[4733]: E0318 10:51:28.176627 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:51:29 crc kubenswrapper[4733]: I0318 10:51:29.176157 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:51:29 crc kubenswrapper[4733]: E0318 10:51:29.176874 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:51:38 crc kubenswrapper[4733]: I0318 10:51:38.176995 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:51:38 crc kubenswrapper[4733]: E0318 10:51:38.178269 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:51:39 crc kubenswrapper[4733]: I0318 10:51:39.176340 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:51:39 crc kubenswrapper[4733]: E0318 10:51:39.177531 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:51:42 crc kubenswrapper[4733]: I0318 10:51:42.175723 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:51:42 crc kubenswrapper[4733]: E0318 10:51:42.176506 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:51:50 crc kubenswrapper[4733]: I0318 10:51:50.176158 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:51:50 crc kubenswrapper[4733]: E0318 10:51:50.177325 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:51:54 crc kubenswrapper[4733]: I0318 10:51:54.175895 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:51:54 crc kubenswrapper[4733]: E0318 10:51:54.176881 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:51:56 crc kubenswrapper[4733]: I0318 10:51:56.176672 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:51:56 crc kubenswrapper[4733]: E0318 10:51:56.177386 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.158322 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563852-nxds9"] Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159608 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159632 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159659 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2166cf23-2a65-4b17-922e-3131be1d6d8b" containerName="oc" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159670 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2166cf23-2a65-4b17-922e-3131be1d6d8b" containerName="oc" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159695 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159708 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159724 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="extract-content" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159734 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="extract-content" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159757 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="extract-content" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159768 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="extract-content" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159796 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="extract-utilities" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159808 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="extract-utilities" Mar 18 10:52:00 crc kubenswrapper[4733]: E0318 10:52:00.159823 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="extract-utilities" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.159832 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="extract-utilities" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.160103 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b49b68f2-ffd0-4ff7-a4ae-691c0a71e896" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.160134 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c6e055-5d64-4158-b0d2-27de0cb9f7c2" containerName="registry-server" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.160155 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2166cf23-2a65-4b17-922e-3131be1d6d8b" containerName="oc" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.160913 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.164219 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563852-nxds9"] Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.165008 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.165901 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.166489 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.287672 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq8r5\" (UniqueName: \"kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5\") pod \"auto-csr-approver-29563852-nxds9\" (UID: \"21264fda-07b1-4a7f-ac61-432c6dc9a230\") " pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.389743 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq8r5\" (UniqueName: \"kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5\") pod \"auto-csr-approver-29563852-nxds9\" (UID: \"21264fda-07b1-4a7f-ac61-432c6dc9a230\") " pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.422632 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq8r5\" (UniqueName: \"kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5\") pod \"auto-csr-approver-29563852-nxds9\" (UID: \"21264fda-07b1-4a7f-ac61-432c6dc9a230\") " pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.486582 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:00 crc kubenswrapper[4733]: I0318 10:52:00.973021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563852-nxds9"] Mar 18 10:52:00 crc kubenswrapper[4733]: W0318 10:52:00.990067 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21264fda_07b1_4a7f_ac61_432c6dc9a230.slice/crio-ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15 WatchSource:0}: Error finding container ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15: Status 404 returned error can't find the container with id ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15 Mar 18 10:52:01 crc kubenswrapper[4733]: I0318 10:52:01.683601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563852-nxds9" event={"ID":"21264fda-07b1-4a7f-ac61-432c6dc9a230","Type":"ContainerStarted","Data":"ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15"} Mar 18 10:52:02 crc kubenswrapper[4733]: I0318 10:52:02.696287 4733 generic.go:334] "Generic (PLEG): container finished" podID="21264fda-07b1-4a7f-ac61-432c6dc9a230" containerID="6234fbb28241739a6b36f7e66aab35dc25489c2ddedad91a0ad07ea33e77be17" exitCode=0 Mar 18 10:52:02 crc kubenswrapper[4733]: I0318 10:52:02.696415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563852-nxds9" event={"ID":"21264fda-07b1-4a7f-ac61-432c6dc9a230","Type":"ContainerDied","Data":"6234fbb28241739a6b36f7e66aab35dc25489c2ddedad91a0ad07ea33e77be17"} Mar 18 10:52:03 crc kubenswrapper[4733]: I0318 10:52:03.179249 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:52:03 crc kubenswrapper[4733]: E0318 10:52:03.180128 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.092353 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.158775 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq8r5\" (UniqueName: \"kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5\") pod \"21264fda-07b1-4a7f-ac61-432c6dc9a230\" (UID: \"21264fda-07b1-4a7f-ac61-432c6dc9a230\") " Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.164988 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5" (OuterVolumeSpecName: "kube-api-access-nq8r5") pod "21264fda-07b1-4a7f-ac61-432c6dc9a230" (UID: "21264fda-07b1-4a7f-ac61-432c6dc9a230"). InnerVolumeSpecName "kube-api-access-nq8r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.261239 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq8r5\" (UniqueName: \"kubernetes.io/projected/21264fda-07b1-4a7f-ac61-432c6dc9a230-kube-api-access-nq8r5\") on node \"crc\" DevicePath \"\"" Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.715990 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563852-nxds9" event={"ID":"21264fda-07b1-4a7f-ac61-432c6dc9a230","Type":"ContainerDied","Data":"ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15"} Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.716045 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff53466daa2cf7b98be28892b619a1f0a9aab3cc5dc1eb983f44b8effab0bf15" Mar 18 10:52:04 crc kubenswrapper[4733]: I0318 10:52:04.716063 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563852-nxds9" Mar 18 10:52:05 crc kubenswrapper[4733]: I0318 10:52:05.177353 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:52:05 crc kubenswrapper[4733]: E0318 10:52:05.177830 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:52:05 crc kubenswrapper[4733]: I0318 10:52:05.217476 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563846-zfzc4"] Mar 18 10:52:05 crc kubenswrapper[4733]: I0318 10:52:05.217549 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563846-zfzc4"] Mar 18 10:52:07 crc kubenswrapper[4733]: I0318 10:52:07.191673 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d9650e-918c-4ff4-82bd-ba01e08b6588" path="/var/lib/kubelet/pods/02d9650e-918c-4ff4-82bd-ba01e08b6588/volumes" Mar 18 10:52:10 crc kubenswrapper[4733]: I0318 10:52:10.847132 4733 scope.go:117] "RemoveContainer" containerID="df6a0cb730f6fbfc1289efd16f5838442dcf854d748f1b313f239fe3a8ed31b9" Mar 18 10:52:11 crc kubenswrapper[4733]: I0318 10:52:11.185240 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:52:11 crc kubenswrapper[4733]: E0318 10:52:11.185681 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:15 crc kubenswrapper[4733]: I0318 10:52:15.175827 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:52:15 crc kubenswrapper[4733]: E0318 10:52:15.176704 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:52:16 crc kubenswrapper[4733]: I0318 10:52:16.176642 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:52:16 crc kubenswrapper[4733]: E0318 10:52:16.176939 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:52:22 crc kubenswrapper[4733]: I0318 10:52:22.176543 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:52:22 crc kubenswrapper[4733]: E0318 10:52:22.177278 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:27 crc kubenswrapper[4733]: I0318 10:52:27.175366 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:52:27 crc kubenswrapper[4733]: E0318 10:52:27.176052 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:52:29 crc kubenswrapper[4733]: I0318 10:52:29.176270 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:52:29 crc kubenswrapper[4733]: E0318 10:52:29.180705 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:52:35 crc kubenswrapper[4733]: I0318 10:52:35.175838 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:52:35 crc kubenswrapper[4733]: E0318 10:52:35.176828 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:40 crc kubenswrapper[4733]: I0318 10:52:40.176257 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:52:40 crc kubenswrapper[4733]: E0318 10:52:40.177393 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:52:40 crc kubenswrapper[4733]: I0318 10:52:40.177393 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:52:40 crc kubenswrapper[4733]: E0318 10:52:40.178246 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:52:46 crc kubenswrapper[4733]: I0318 10:52:46.176473 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:52:46 crc kubenswrapper[4733]: E0318 10:52:46.177305 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.089869 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:52:48 crc kubenswrapper[4733]: E0318 10:52:48.090382 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21264fda-07b1-4a7f-ac61-432c6dc9a230" containerName="oc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.090405 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="21264fda-07b1-4a7f-ac61-432c6dc9a230" containerName="oc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.090732 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="21264fda-07b1-4a7f-ac61-432c6dc9a230" containerName="oc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.092669 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.124781 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.198483 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r46zw\" (UniqueName: \"kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.198619 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.198705 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.300109 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.300359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.300446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r46zw\" (UniqueName: \"kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.300699 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.300881 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.323282 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r46zw\" (UniqueName: \"kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw\") pod \"certified-operators-2vhhc\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.439549 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:48 crc kubenswrapper[4733]: I0318 10:52:48.718013 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:52:49 crc kubenswrapper[4733]: I0318 10:52:49.140963 4733 generic.go:334] "Generic (PLEG): container finished" podID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerID="0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d" exitCode=0 Mar 18 10:52:49 crc kubenswrapper[4733]: I0318 10:52:49.141077 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerDied","Data":"0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d"} Mar 18 10:52:49 crc kubenswrapper[4733]: I0318 10:52:49.141548 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerStarted","Data":"bfb7ae10f9639c8b35d73989ab0d314c8fd91b3fa87651193e323db772fbad71"} Mar 18 10:52:50 crc kubenswrapper[4733]: I0318 10:52:50.150166 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerStarted","Data":"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50"} Mar 18 10:52:51 crc kubenswrapper[4733]: I0318 10:52:51.164113 4733 generic.go:334] "Generic (PLEG): container finished" podID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerID="4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50" exitCode=0 Mar 18 10:52:51 crc kubenswrapper[4733]: I0318 10:52:51.164178 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerDied","Data":"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50"} Mar 18 10:52:52 crc kubenswrapper[4733]: I0318 10:52:52.176103 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:52:52 crc kubenswrapper[4733]: E0318 10:52:52.176942 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:52:52 crc kubenswrapper[4733]: I0318 10:52:52.179616 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerStarted","Data":"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738"} Mar 18 10:52:52 crc kubenswrapper[4733]: I0318 10:52:52.204782 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2vhhc" podStartSLOduration=1.801296522 podStartE2EDuration="4.204763189s" podCreationTimestamp="2026-03-18 10:52:48 +0000 UTC" firstStartedPulling="2026-03-18 10:52:49.143559419 +0000 UTC m=+2408.635293774" lastFinishedPulling="2026-03-18 10:52:51.547026116 +0000 UTC m=+2411.038760441" observedRunningTime="2026-03-18 10:52:52.199883931 +0000 UTC m=+2411.691618256" watchObservedRunningTime="2026-03-18 10:52:52.204763189 +0000 UTC m=+2411.696497514" Mar 18 10:52:53 crc kubenswrapper[4733]: I0318 10:52:53.176417 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:52:53 crc kubenswrapper[4733]: E0318 10:52:53.177042 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:52:58 crc kubenswrapper[4733]: I0318 10:52:58.175786 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:52:58 crc kubenswrapper[4733]: E0318 10:52:58.177180 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:52:58 crc kubenswrapper[4733]: I0318 10:52:58.439776 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:58 crc kubenswrapper[4733]: I0318 10:52:58.439857 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:58 crc kubenswrapper[4733]: I0318 10:52:58.504563 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:59 crc kubenswrapper[4733]: I0318 10:52:59.324037 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:52:59 crc kubenswrapper[4733]: I0318 10:52:59.383507 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.269077 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2vhhc" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="registry-server" containerID="cri-o://f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738" gracePeriod=2 Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.764157 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.834262 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content\") pod \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.834445 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities\") pod \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.834545 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r46zw\" (UniqueName: \"kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw\") pod \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\" (UID: \"c884f6b7-6551-4a77-b19f-d0ea8c634eb2\") " Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.835519 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities" (OuterVolumeSpecName: "utilities") pod "c884f6b7-6551-4a77-b19f-d0ea8c634eb2" (UID: "c884f6b7-6551-4a77-b19f-d0ea8c634eb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.843697 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw" (OuterVolumeSpecName: "kube-api-access-r46zw") pod "c884f6b7-6551-4a77-b19f-d0ea8c634eb2" (UID: "c884f6b7-6551-4a77-b19f-d0ea8c634eb2"). InnerVolumeSpecName "kube-api-access-r46zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.910375 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c884f6b7-6551-4a77-b19f-d0ea8c634eb2" (UID: "c884f6b7-6551-4a77-b19f-d0ea8c634eb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.936057 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r46zw\" (UniqueName: \"kubernetes.io/projected/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-kube-api-access-r46zw\") on node \"crc\" DevicePath \"\"" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.936099 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:53:01 crc kubenswrapper[4733]: I0318 10:53:01.936110 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c884f6b7-6551-4a77-b19f-d0ea8c634eb2-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.280642 4733 generic.go:334] "Generic (PLEG): container finished" podID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerID="f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738" exitCode=0 Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.280720 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vhhc" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.280701 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerDied","Data":"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738"} Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.281231 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vhhc" event={"ID":"c884f6b7-6551-4a77-b19f-d0ea8c634eb2","Type":"ContainerDied","Data":"bfb7ae10f9639c8b35d73989ab0d314c8fd91b3fa87651193e323db772fbad71"} Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.281269 4733 scope.go:117] "RemoveContainer" containerID="f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.322220 4733 scope.go:117] "RemoveContainer" containerID="4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.329867 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.349673 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2vhhc"] Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.360266 4733 scope.go:117] "RemoveContainer" containerID="0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.385667 4733 scope.go:117] "RemoveContainer" containerID="f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738" Mar 18 10:53:02 crc kubenswrapper[4733]: E0318 10:53:02.386075 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738\": container with ID starting with f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738 not found: ID does not exist" containerID="f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.386110 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738"} err="failed to get container status \"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738\": rpc error: code = NotFound desc = could not find container \"f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738\": container with ID starting with f289f6deea9d24ec4de7af790bd1a748a15b66a6de7ff132e8b3a144800ff738 not found: ID does not exist" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.386135 4733 scope.go:117] "RemoveContainer" containerID="4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50" Mar 18 10:53:02 crc kubenswrapper[4733]: E0318 10:53:02.386565 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50\": container with ID starting with 4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50 not found: ID does not exist" containerID="4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.386589 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50"} err="failed to get container status \"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50\": rpc error: code = NotFound desc = could not find container \"4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50\": container with ID starting with 4f56f4c122771f35159f24bcb1a971cd9aaf2bdd30567d91aef109eb055c2b50 not found: ID does not exist" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.386607 4733 scope.go:117] "RemoveContainer" containerID="0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d" Mar 18 10:53:02 crc kubenswrapper[4733]: E0318 10:53:02.386841 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d\": container with ID starting with 0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d not found: ID does not exist" containerID="0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d" Mar 18 10:53:02 crc kubenswrapper[4733]: I0318 10:53:02.386869 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d"} err="failed to get container status \"0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d\": rpc error: code = NotFound desc = could not find container \"0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d\": container with ID starting with 0a9f0b6652df7606f5074c262197297d4120c34ea16c27776883a78daecb640d not found: ID does not exist" Mar 18 10:53:03 crc kubenswrapper[4733]: I0318 10:53:03.197808 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" path="/var/lib/kubelet/pods/c884f6b7-6551-4a77-b19f-d0ea8c634eb2/volumes" Mar 18 10:53:04 crc kubenswrapper[4733]: I0318 10:53:04.176007 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:53:04 crc kubenswrapper[4733]: E0318 10:53:04.176534 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:53:07 crc kubenswrapper[4733]: I0318 10:53:07.175463 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:53:07 crc kubenswrapper[4733]: E0318 10:53:07.176461 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:53:09 crc kubenswrapper[4733]: I0318 10:53:09.176561 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:53:09 crc kubenswrapper[4733]: E0318 10:53:09.177393 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:53:18 crc kubenswrapper[4733]: I0318 10:53:18.176200 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:53:18 crc kubenswrapper[4733]: I0318 10:53:18.177141 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:53:18 crc kubenswrapper[4733]: E0318 10:53:18.177433 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:53:18 crc kubenswrapper[4733]: E0318 10:53:18.177669 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:53:23 crc kubenswrapper[4733]: I0318 10:53:23.176111 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:53:23 crc kubenswrapper[4733]: E0318 10:53:23.177277 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:53:29 crc kubenswrapper[4733]: I0318 10:53:29.179286 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:53:29 crc kubenswrapper[4733]: E0318 10:53:29.181585 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:53:31 crc kubenswrapper[4733]: I0318 10:53:31.181657 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:53:31 crc kubenswrapper[4733]: I0318 10:53:31.577765 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187"} Mar 18 10:53:31 crc kubenswrapper[4733]: I0318 10:53:31.578876 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:53:34 crc kubenswrapper[4733]: I0318 10:53:34.175766 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:53:34 crc kubenswrapper[4733]: E0318 10:53:34.176513 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:53:36 crc kubenswrapper[4733]: I0318 10:53:36.633272 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" exitCode=0 Mar 18 10:53:36 crc kubenswrapper[4733]: I0318 10:53:36.633332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187"} Mar 18 10:53:36 crc kubenswrapper[4733]: I0318 10:53:36.634558 4733 scope.go:117] "RemoveContainer" containerID="52c377e9a60c9ca96c08e610b060d46b5fcfe8f4ca8351f71d96116255ccee60" Mar 18 10:53:36 crc kubenswrapper[4733]: I0318 10:53:36.635628 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:53:36 crc kubenswrapper[4733]: E0318 10:53:36.636042 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:53:44 crc kubenswrapper[4733]: I0318 10:53:44.175063 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:53:44 crc kubenswrapper[4733]: E0318 10:53:44.177258 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:53:45 crc kubenswrapper[4733]: I0318 10:53:45.176498 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:53:45 crc kubenswrapper[4733]: I0318 10:53:45.727655 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c"} Mar 18 10:53:45 crc kubenswrapper[4733]: I0318 10:53:45.728177 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:53:49 crc kubenswrapper[4733]: I0318 10:53:49.783158 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" exitCode=0 Mar 18 10:53:49 crc kubenswrapper[4733]: I0318 10:53:49.783330 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c"} Mar 18 10:53:49 crc kubenswrapper[4733]: I0318 10:53:49.784075 4733 scope.go:117] "RemoveContainer" containerID="615d6075a16d1723238d5f484c97fecdaa694488b8494a55c0a3d329cf030b8f" Mar 18 10:53:49 crc kubenswrapper[4733]: I0318 10:53:49.785513 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:53:49 crc kubenswrapper[4733]: E0318 10:53:49.786622 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:53:50 crc kubenswrapper[4733]: I0318 10:53:50.176177 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:53:50 crc kubenswrapper[4733]: E0318 10:53:50.176624 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:53:59 crc kubenswrapper[4733]: I0318 10:53:59.176267 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:53:59 crc kubenswrapper[4733]: E0318 10:53:59.177170 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.167053 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563854-8xm7m"] Mar 18 10:54:00 crc kubenswrapper[4733]: E0318 10:54:00.167763 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="extract-utilities" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.167808 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="extract-utilities" Mar 18 10:54:00 crc kubenswrapper[4733]: E0318 10:54:00.167876 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="registry-server" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.167896 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="registry-server" Mar 18 10:54:00 crc kubenswrapper[4733]: E0318 10:54:00.167937 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="extract-content" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.167955 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="extract-content" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.168360 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c884f6b7-6551-4a77-b19f-d0ea8c634eb2" containerName="registry-server" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.169295 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.174417 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.174434 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.177244 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.190026 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563854-8xm7m"] Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.246081 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zl8r\" (UniqueName: \"kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r\") pod \"auto-csr-approver-29563854-8xm7m\" (UID: \"e2194f8e-8219-4d20-9657-bab035e9ce0b\") " pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.347607 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zl8r\" (UniqueName: \"kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r\") pod \"auto-csr-approver-29563854-8xm7m\" (UID: \"e2194f8e-8219-4d20-9657-bab035e9ce0b\") " pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.386633 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zl8r\" (UniqueName: \"kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r\") pod \"auto-csr-approver-29563854-8xm7m\" (UID: \"e2194f8e-8219-4d20-9657-bab035e9ce0b\") " pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.498402 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:00 crc kubenswrapper[4733]: I0318 10:54:00.981816 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563854-8xm7m"] Mar 18 10:54:00 crc kubenswrapper[4733]: W0318 10:54:00.982988 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2194f8e_8219_4d20_9657_bab035e9ce0b.slice/crio-b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d WatchSource:0}: Error finding container b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d: Status 404 returned error can't find the container with id b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d Mar 18 10:54:01 crc kubenswrapper[4733]: I0318 10:54:01.948557 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" event={"ID":"e2194f8e-8219-4d20-9657-bab035e9ce0b","Type":"ContainerStarted","Data":"b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d"} Mar 18 10:54:02 crc kubenswrapper[4733]: I0318 10:54:02.959712 4733 generic.go:334] "Generic (PLEG): container finished" podID="e2194f8e-8219-4d20-9657-bab035e9ce0b" containerID="0646c2eb1d4076069ba17429b100767c9ea92208b7525c26c0789773916b849f" exitCode=0 Mar 18 10:54:02 crc kubenswrapper[4733]: I0318 10:54:02.959815 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" event={"ID":"e2194f8e-8219-4d20-9657-bab035e9ce0b","Type":"ContainerDied","Data":"0646c2eb1d4076069ba17429b100767c9ea92208b7525c26c0789773916b849f"} Mar 18 10:54:03 crc kubenswrapper[4733]: I0318 10:54:03.176134 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:54:03 crc kubenswrapper[4733]: E0318 10:54:03.177050 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.313230 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.335248 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zl8r\" (UniqueName: \"kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r\") pod \"e2194f8e-8219-4d20-9657-bab035e9ce0b\" (UID: \"e2194f8e-8219-4d20-9657-bab035e9ce0b\") " Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.346582 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r" (OuterVolumeSpecName: "kube-api-access-5zl8r") pod "e2194f8e-8219-4d20-9657-bab035e9ce0b" (UID: "e2194f8e-8219-4d20-9657-bab035e9ce0b"). InnerVolumeSpecName "kube-api-access-5zl8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.437077 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zl8r\" (UniqueName: \"kubernetes.io/projected/e2194f8e-8219-4d20-9657-bab035e9ce0b-kube-api-access-5zl8r\") on node \"crc\" DevicePath \"\"" Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.979346 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" event={"ID":"e2194f8e-8219-4d20-9657-bab035e9ce0b","Type":"ContainerDied","Data":"b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d"} Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.979387 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b460879da7c6f247784fa731b833b199cebd1e8920e884c338530badccb40f7d" Mar 18 10:54:04 crc kubenswrapper[4733]: I0318 10:54:04.979425 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563854-8xm7m" Mar 18 10:54:05 crc kubenswrapper[4733]: I0318 10:54:05.175602 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:54:05 crc kubenswrapper[4733]: E0318 10:54:05.175821 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:54:05 crc kubenswrapper[4733]: I0318 10:54:05.399558 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563848-2tfjq"] Mar 18 10:54:05 crc kubenswrapper[4733]: I0318 10:54:05.405700 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563848-2tfjq"] Mar 18 10:54:07 crc kubenswrapper[4733]: I0318 10:54:07.186311 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc1a1de-2965-422f-8ac4-77358d0d7df6" path="/var/lib/kubelet/pods/1bc1a1de-2965-422f-8ac4-77358d0d7df6/volumes" Mar 18 10:54:10 crc kubenswrapper[4733]: I0318 10:54:10.987870 4733 scope.go:117] "RemoveContainer" containerID="6fe3d8a40e1ba17924153a168a35cf8fd5e9cc3d1fdeb9b0f70b81b8350f5f56" Mar 18 10:54:14 crc kubenswrapper[4733]: I0318 10:54:14.175853 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:54:14 crc kubenswrapper[4733]: E0318 10:54:14.176860 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:54:14 crc kubenswrapper[4733]: I0318 10:54:14.177872 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:54:14 crc kubenswrapper[4733]: E0318 10:54:14.178243 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:54:17 crc kubenswrapper[4733]: I0318 10:54:17.175380 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:54:17 crc kubenswrapper[4733]: E0318 10:54:17.177233 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:54:25 crc kubenswrapper[4733]: I0318 10:54:25.175420 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:54:25 crc kubenswrapper[4733]: E0318 10:54:25.176611 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:54:27 crc kubenswrapper[4733]: I0318 10:54:27.175987 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:54:27 crc kubenswrapper[4733]: E0318 10:54:27.176425 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:54:32 crc kubenswrapper[4733]: I0318 10:54:32.176571 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:54:32 crc kubenswrapper[4733]: E0318 10:54:32.177566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:54:37 crc kubenswrapper[4733]: I0318 10:54:37.176092 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:54:37 crc kubenswrapper[4733]: E0318 10:54:37.177363 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:54:41 crc kubenswrapper[4733]: I0318 10:54:41.182358 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:54:41 crc kubenswrapper[4733]: E0318 10:54:41.182892 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:54:44 crc kubenswrapper[4733]: I0318 10:54:44.176296 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:54:44 crc kubenswrapper[4733]: E0318 10:54:44.177319 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:54:48 crc kubenswrapper[4733]: I0318 10:54:48.176002 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:54:48 crc kubenswrapper[4733]: E0318 10:54:48.176711 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:54:55 crc kubenswrapper[4733]: I0318 10:54:55.175542 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:54:55 crc kubenswrapper[4733]: E0318 10:54:55.176545 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:54:58 crc kubenswrapper[4733]: I0318 10:54:58.176977 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:54:58 crc kubenswrapper[4733]: E0318 10:54:58.177997 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:55:02 crc kubenswrapper[4733]: I0318 10:55:02.176794 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:55:02 crc kubenswrapper[4733]: E0318 10:55:02.177696 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:55:10 crc kubenswrapper[4733]: I0318 10:55:10.175642 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:55:10 crc kubenswrapper[4733]: E0318 10:55:10.178133 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:55:11 crc kubenswrapper[4733]: I0318 10:55:11.180286 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:55:11 crc kubenswrapper[4733]: E0318 10:55:11.181009 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:55:13 crc kubenswrapper[4733]: I0318 10:55:13.176085 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:55:13 crc kubenswrapper[4733]: E0318 10:55:13.176537 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:55:25 crc kubenswrapper[4733]: I0318 10:55:25.177017 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:55:25 crc kubenswrapper[4733]: E0318 10:55:25.178029 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:55:25 crc kubenswrapper[4733]: I0318 10:55:25.178223 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:55:25 crc kubenswrapper[4733]: E0318 10:55:25.178783 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:55:27 crc kubenswrapper[4733]: I0318 10:55:27.175485 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:55:27 crc kubenswrapper[4733]: E0318 10:55:27.176290 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:55:36 crc kubenswrapper[4733]: I0318 10:55:36.176585 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:55:36 crc kubenswrapper[4733]: E0318 10:55:36.177713 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:55:38 crc kubenswrapper[4733]: I0318 10:55:38.177311 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:55:38 crc kubenswrapper[4733]: E0318 10:55:38.178674 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:55:39 crc kubenswrapper[4733]: I0318 10:55:39.176475 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:55:39 crc kubenswrapper[4733]: E0318 10:55:39.177456 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:55:51 crc kubenswrapper[4733]: I0318 10:55:51.185110 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:55:51 crc kubenswrapper[4733]: E0318 10:55:51.187689 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:55:52 crc kubenswrapper[4733]: I0318 10:55:52.175664 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:55:52 crc kubenswrapper[4733]: E0318 10:55:52.176095 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:55:54 crc kubenswrapper[4733]: I0318 10:55:54.175655 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:55:54 crc kubenswrapper[4733]: E0318 10:55:54.176305 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.154624 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563856-qhf5l"] Mar 18 10:56:00 crc kubenswrapper[4733]: E0318 10:56:00.155067 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2194f8e-8219-4d20-9657-bab035e9ce0b" containerName="oc" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.155087 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2194f8e-8219-4d20-9657-bab035e9ce0b" containerName="oc" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.155428 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2194f8e-8219-4d20-9657-bab035e9ce0b" containerName="oc" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.156336 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.161958 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.162835 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.163752 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.169937 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563856-qhf5l"] Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.226037 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhtd\" (UniqueName: \"kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd\") pod \"auto-csr-approver-29563856-qhf5l\" (UID: \"e1a4900e-15be-4e3f-a8a2-3eb582acbc20\") " pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.328026 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhtd\" (UniqueName: \"kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd\") pod \"auto-csr-approver-29563856-qhf5l\" (UID: \"e1a4900e-15be-4e3f-a8a2-3eb582acbc20\") " pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.353943 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhtd\" (UniqueName: \"kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd\") pod \"auto-csr-approver-29563856-qhf5l\" (UID: \"e1a4900e-15be-4e3f-a8a2-3eb582acbc20\") " pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.518792 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.805171 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563856-qhf5l"] Mar 18 10:56:00 crc kubenswrapper[4733]: W0318 10:56:00.807728 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1a4900e_15be_4e3f_a8a2_3eb582acbc20.slice/crio-74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e WatchSource:0}: Error finding container 74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e: Status 404 returned error can't find the container with id 74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e Mar 18 10:56:00 crc kubenswrapper[4733]: I0318 10:56:00.810676 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 10:56:01 crc kubenswrapper[4733]: I0318 10:56:01.156725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" event={"ID":"e1a4900e-15be-4e3f-a8a2-3eb582acbc20","Type":"ContainerStarted","Data":"74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e"} Mar 18 10:56:02 crc kubenswrapper[4733]: I0318 10:56:02.165715 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" event={"ID":"e1a4900e-15be-4e3f-a8a2-3eb582acbc20","Type":"ContainerStarted","Data":"9c09a3cb9db31583aa867b9b2e7873c25af33c3cb06cde66bc60959b1e039850"} Mar 18 10:56:02 crc kubenswrapper[4733]: I0318 10:56:02.187684 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" podStartSLOduration=1.223335041 podStartE2EDuration="2.187665025s" podCreationTimestamp="2026-03-18 10:56:00 +0000 UTC" firstStartedPulling="2026-03-18 10:56:00.810167164 +0000 UTC m=+2600.301901519" lastFinishedPulling="2026-03-18 10:56:01.774497148 +0000 UTC m=+2601.266231503" observedRunningTime="2026-03-18 10:56:02.178024194 +0000 UTC m=+2601.669758529" watchObservedRunningTime="2026-03-18 10:56:02.187665025 +0000 UTC m=+2601.679399360" Mar 18 10:56:02 crc kubenswrapper[4733]: E0318 10:56:02.533777 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1a4900e_15be_4e3f_a8a2_3eb582acbc20.slice/crio-conmon-9c09a3cb9db31583aa867b9b2e7873c25af33c3cb06cde66bc60959b1e039850.scope\": RecentStats: unable to find data in memory cache]" Mar 18 10:56:03 crc kubenswrapper[4733]: I0318 10:56:03.176890 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:56:03 crc kubenswrapper[4733]: E0318 10:56:03.180811 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 10:56:03 crc kubenswrapper[4733]: I0318 10:56:03.185532 4733 generic.go:334] "Generic (PLEG): container finished" podID="e1a4900e-15be-4e3f-a8a2-3eb582acbc20" containerID="9c09a3cb9db31583aa867b9b2e7873c25af33c3cb06cde66bc60959b1e039850" exitCode=0 Mar 18 10:56:03 crc kubenswrapper[4733]: I0318 10:56:03.203267 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" event={"ID":"e1a4900e-15be-4e3f-a8a2-3eb582acbc20","Type":"ContainerDied","Data":"9c09a3cb9db31583aa867b9b2e7873c25af33c3cb06cde66bc60959b1e039850"} Mar 18 10:56:04 crc kubenswrapper[4733]: I0318 10:56:04.533377 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:04 crc kubenswrapper[4733]: I0318 10:56:04.709640 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmhtd\" (UniqueName: \"kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd\") pod \"e1a4900e-15be-4e3f-a8a2-3eb582acbc20\" (UID: \"e1a4900e-15be-4e3f-a8a2-3eb582acbc20\") " Mar 18 10:56:04 crc kubenswrapper[4733]: I0318 10:56:04.719777 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd" (OuterVolumeSpecName: "kube-api-access-vmhtd") pod "e1a4900e-15be-4e3f-a8a2-3eb582acbc20" (UID: "e1a4900e-15be-4e3f-a8a2-3eb582acbc20"). InnerVolumeSpecName "kube-api-access-vmhtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:56:04 crc kubenswrapper[4733]: I0318 10:56:04.812026 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmhtd\" (UniqueName: \"kubernetes.io/projected/e1a4900e-15be-4e3f-a8a2-3eb582acbc20-kube-api-access-vmhtd\") on node \"crc\" DevicePath \"\"" Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.176416 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:56:05 crc kubenswrapper[4733]: E0318 10:56:05.176781 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.207614 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" event={"ID":"e1a4900e-15be-4e3f-a8a2-3eb582acbc20","Type":"ContainerDied","Data":"74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e"} Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.207661 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74c5ad720d80d58ad8720f159874e34d6dba44d12a6a557c0ec232278f2fae2e" Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.207690 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563856-qhf5l" Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.620367 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563850-8gwpw"] Mar 18 10:56:05 crc kubenswrapper[4733]: I0318 10:56:05.631876 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563850-8gwpw"] Mar 18 10:56:07 crc kubenswrapper[4733]: I0318 10:56:07.186267 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2166cf23-2a65-4b17-922e-3131be1d6d8b" path="/var/lib/kubelet/pods/2166cf23-2a65-4b17-922e-3131be1d6d8b/volumes" Mar 18 10:56:10 crc kubenswrapper[4733]: I0318 10:56:10.114853 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:56:10 crc kubenswrapper[4733]: E0318 10:56:10.118835 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:56:11 crc kubenswrapper[4733]: I0318 10:56:11.107843 4733 scope.go:117] "RemoveContainer" containerID="649f36c8155821a228e5fee55c54c1d5edbde655cee7563ac249384dedef675b" Mar 18 10:56:15 crc kubenswrapper[4733]: I0318 10:56:15.275464 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:56:16 crc kubenswrapper[4733]: I0318 10:56:16.175998 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:56:16 crc kubenswrapper[4733]: E0318 10:56:16.177420 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:56:16 crc kubenswrapper[4733]: I0318 10:56:16.317070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56"} Mar 18 10:56:21 crc kubenswrapper[4733]: I0318 10:56:21.183828 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:56:21 crc kubenswrapper[4733]: E0318 10:56:21.185022 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:56:31 crc kubenswrapper[4733]: I0318 10:56:31.184477 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:56:31 crc kubenswrapper[4733]: E0318 10:56:31.185539 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:56:34 crc kubenswrapper[4733]: I0318 10:56:34.175666 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:56:34 crc kubenswrapper[4733]: E0318 10:56:34.176171 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:56:45 crc kubenswrapper[4733]: I0318 10:56:45.176084 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:56:45 crc kubenswrapper[4733]: E0318 10:56:45.177409 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:56:49 crc kubenswrapper[4733]: I0318 10:56:49.176556 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:56:49 crc kubenswrapper[4733]: E0318 10:56:49.177613 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:56:58 crc kubenswrapper[4733]: I0318 10:56:58.176347 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:56:58 crc kubenswrapper[4733]: E0318 10:56:58.177261 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:57:03 crc kubenswrapper[4733]: I0318 10:57:03.176680 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:57:03 crc kubenswrapper[4733]: E0318 10:57:03.177736 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:57:09 crc kubenswrapper[4733]: I0318 10:57:09.176907 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:57:09 crc kubenswrapper[4733]: E0318 10:57:09.177911 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:57:18 crc kubenswrapper[4733]: I0318 10:57:18.175504 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:57:18 crc kubenswrapper[4733]: E0318 10:57:18.176469 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:57:22 crc kubenswrapper[4733]: I0318 10:57:22.176861 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:57:22 crc kubenswrapper[4733]: E0318 10:57:22.177591 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:57:30 crc kubenswrapper[4733]: I0318 10:57:30.175606 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:57:30 crc kubenswrapper[4733]: E0318 10:57:30.176234 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:57:34 crc kubenswrapper[4733]: I0318 10:57:34.175959 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:57:34 crc kubenswrapper[4733]: E0318 10:57:34.176825 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:57:45 crc kubenswrapper[4733]: I0318 10:57:45.175910 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:57:45 crc kubenswrapper[4733]: E0318 10:57:45.176949 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:57:49 crc kubenswrapper[4733]: I0318 10:57:49.176156 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:57:49 crc kubenswrapper[4733]: E0318 10:57:49.176782 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:57:58 crc kubenswrapper[4733]: I0318 10:57:58.176016 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:57:58 crc kubenswrapper[4733]: E0318 10:57:58.177101 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.152415 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563858-bxq6r"] Mar 18 10:58:00 crc kubenswrapper[4733]: E0318 10:58:00.152916 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1a4900e-15be-4e3f-a8a2-3eb582acbc20" containerName="oc" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.152940 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1a4900e-15be-4e3f-a8a2-3eb582acbc20" containerName="oc" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.153345 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1a4900e-15be-4e3f-a8a2-3eb582acbc20" containerName="oc" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.154138 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.157531 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.157890 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.159097 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.164718 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563858-bxq6r"] Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.347564 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9pdm\" (UniqueName: \"kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm\") pod \"auto-csr-approver-29563858-bxq6r\" (UID: \"7bc61f2d-1837-4253-a3a3-91d8acc950f8\") " pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.449476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9pdm\" (UniqueName: \"kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm\") pod \"auto-csr-approver-29563858-bxq6r\" (UID: \"7bc61f2d-1837-4253-a3a3-91d8acc950f8\") " pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.486904 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9pdm\" (UniqueName: \"kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm\") pod \"auto-csr-approver-29563858-bxq6r\" (UID: \"7bc61f2d-1837-4253-a3a3-91d8acc950f8\") " pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:00 crc kubenswrapper[4733]: I0318 10:58:00.780547 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:01 crc kubenswrapper[4733]: I0318 10:58:01.182595 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:01 crc kubenswrapper[4733]: E0318 10:58:01.183616 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:58:01 crc kubenswrapper[4733]: I0318 10:58:01.336284 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563858-bxq6r"] Mar 18 10:58:01 crc kubenswrapper[4733]: I0318 10:58:01.551505 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" event={"ID":"7bc61f2d-1837-4253-a3a3-91d8acc950f8","Type":"ContainerStarted","Data":"572677742ca6f2dd1a03ef173b7a9bb884073e34767cd09079a938a5f114183a"} Mar 18 10:58:03 crc kubenswrapper[4733]: I0318 10:58:03.571515 4733 generic.go:334] "Generic (PLEG): container finished" podID="7bc61f2d-1837-4253-a3a3-91d8acc950f8" containerID="ae5d77ede52fa11bd913773d4add1f40cd6fcaf6154c4236eccd984879ea57ff" exitCode=0 Mar 18 10:58:03 crc kubenswrapper[4733]: I0318 10:58:03.571611 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" event={"ID":"7bc61f2d-1837-4253-a3a3-91d8acc950f8","Type":"ContainerDied","Data":"ae5d77ede52fa11bd913773d4add1f40cd6fcaf6154c4236eccd984879ea57ff"} Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.013042 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.140444 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9pdm\" (UniqueName: \"kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm\") pod \"7bc61f2d-1837-4253-a3a3-91d8acc950f8\" (UID: \"7bc61f2d-1837-4253-a3a3-91d8acc950f8\") " Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.147245 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm" (OuterVolumeSpecName: "kube-api-access-h9pdm") pod "7bc61f2d-1837-4253-a3a3-91d8acc950f8" (UID: "7bc61f2d-1837-4253-a3a3-91d8acc950f8"). InnerVolumeSpecName "kube-api-access-h9pdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.243077 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9pdm\" (UniqueName: \"kubernetes.io/projected/7bc61f2d-1837-4253-a3a3-91d8acc950f8-kube-api-access-h9pdm\") on node \"crc\" DevicePath \"\"" Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.607396 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" event={"ID":"7bc61f2d-1837-4253-a3a3-91d8acc950f8","Type":"ContainerDied","Data":"572677742ca6f2dd1a03ef173b7a9bb884073e34767cd09079a938a5f114183a"} Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.607713 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="572677742ca6f2dd1a03ef173b7a9bb884073e34767cd09079a938a5f114183a" Mar 18 10:58:05 crc kubenswrapper[4733]: I0318 10:58:05.607641 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563858-bxq6r" Mar 18 10:58:06 crc kubenswrapper[4733]: I0318 10:58:06.111459 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563852-nxds9"] Mar 18 10:58:06 crc kubenswrapper[4733]: I0318 10:58:06.121554 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563852-nxds9"] Mar 18 10:58:07 crc kubenswrapper[4733]: I0318 10:58:07.192592 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21264fda-07b1-4a7f-ac61-432c6dc9a230" path="/var/lib/kubelet/pods/21264fda-07b1-4a7f-ac61-432c6dc9a230/volumes" Mar 18 10:58:11 crc kubenswrapper[4733]: I0318 10:58:11.204407 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:58:11 crc kubenswrapper[4733]: E0318 10:58:11.205357 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:58:11 crc kubenswrapper[4733]: I0318 10:58:11.235120 4733 scope.go:117] "RemoveContainer" containerID="6234fbb28241739a6b36f7e66aab35dc25489c2ddedad91a0ad07ea33e77be17" Mar 18 10:58:14 crc kubenswrapper[4733]: I0318 10:58:14.175844 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:14 crc kubenswrapper[4733]: E0318 10:58:14.176666 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:58:23 crc kubenswrapper[4733]: I0318 10:58:23.176116 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:58:23 crc kubenswrapper[4733]: E0318 10:58:23.177109 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:58:27 crc kubenswrapper[4733]: I0318 10:58:27.175751 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:27 crc kubenswrapper[4733]: E0318 10:58:27.176216 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:58:35 crc kubenswrapper[4733]: I0318 10:58:35.178311 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:58:35 crc kubenswrapper[4733]: E0318 10:58:35.179333 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:58:41 crc kubenswrapper[4733]: I0318 10:58:41.188650 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:41 crc kubenswrapper[4733]: E0318 10:58:41.189879 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:58:43 crc kubenswrapper[4733]: I0318 10:58:43.571572 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:58:43 crc kubenswrapper[4733]: I0318 10:58:43.571990 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.195804 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:58:47 crc kubenswrapper[4733]: E0318 10:58:47.196717 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc61f2d-1837-4253-a3a3-91d8acc950f8" containerName="oc" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.196739 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc61f2d-1837-4253-a3a3-91d8acc950f8" containerName="oc" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.197102 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc61f2d-1837-4253-a3a3-91d8acc950f8" containerName="oc" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.209934 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.219148 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.343762 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.344155 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4v5l\" (UniqueName: \"kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.344352 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.445978 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.446170 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4v5l\" (UniqueName: \"kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.446303 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.446786 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.446854 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.482235 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4v5l\" (UniqueName: \"kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l\") pod \"community-operators-dpxzp\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:47 crc kubenswrapper[4733]: I0318 10:58:47.563472 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:48 crc kubenswrapper[4733]: I0318 10:58:48.135447 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:58:48 crc kubenswrapper[4733]: W0318 10:58:48.145417 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2b9d483_b743_44ab_bf0b_f0b22f133576.slice/crio-5d45d05271d7475d8f987b38020b7a838549e3626fa27dfe3061a1df224ec8fe WatchSource:0}: Error finding container 5d45d05271d7475d8f987b38020b7a838549e3626fa27dfe3061a1df224ec8fe: Status 404 returned error can't find the container with id 5d45d05271d7475d8f987b38020b7a838549e3626fa27dfe3061a1df224ec8fe Mar 18 10:58:48 crc kubenswrapper[4733]: I0318 10:58:48.175973 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:58:49 crc kubenswrapper[4733]: I0318 10:58:49.026403 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f"} Mar 18 10:58:49 crc kubenswrapper[4733]: I0318 10:58:49.026998 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 10:58:49 crc kubenswrapper[4733]: I0318 10:58:49.031061 4733 generic.go:334] "Generic (PLEG): container finished" podID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerID="d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680" exitCode=0 Mar 18 10:58:49 crc kubenswrapper[4733]: I0318 10:58:49.031091 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerDied","Data":"d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680"} Mar 18 10:58:49 crc kubenswrapper[4733]: I0318 10:58:49.031133 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerStarted","Data":"5d45d05271d7475d8f987b38020b7a838549e3626fa27dfe3061a1df224ec8fe"} Mar 18 10:58:51 crc kubenswrapper[4733]: I0318 10:58:51.057508 4733 generic.go:334] "Generic (PLEG): container finished" podID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerID="00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9" exitCode=0 Mar 18 10:58:51 crc kubenswrapper[4733]: I0318 10:58:51.058598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerDied","Data":"00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9"} Mar 18 10:58:52 crc kubenswrapper[4733]: I0318 10:58:52.073454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerStarted","Data":"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16"} Mar 18 10:58:52 crc kubenswrapper[4733]: I0318 10:58:52.115177 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dpxzp" podStartSLOduration=2.62362291 podStartE2EDuration="5.115150355s" podCreationTimestamp="2026-03-18 10:58:47 +0000 UTC" firstStartedPulling="2026-03-18 10:58:49.032406064 +0000 UTC m=+2768.524140419" lastFinishedPulling="2026-03-18 10:58:51.523933499 +0000 UTC m=+2771.015667864" observedRunningTime="2026-03-18 10:58:52.101786559 +0000 UTC m=+2771.593520934" watchObservedRunningTime="2026-03-18 10:58:52.115150355 +0000 UTC m=+2771.606884710" Mar 18 10:58:53 crc kubenswrapper[4733]: I0318 10:58:53.088906 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" exitCode=0 Mar 18 10:58:53 crc kubenswrapper[4733]: I0318 10:58:53.088969 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f"} Mar 18 10:58:53 crc kubenswrapper[4733]: I0318 10:58:53.089507 4733 scope.go:117] "RemoveContainer" containerID="ce9239548d170b75a5ed09a485a37b22892a226ac679cb077b3445d3c2f2c187" Mar 18 10:58:53 crc kubenswrapper[4733]: I0318 10:58:53.091076 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:58:53 crc kubenswrapper[4733]: E0318 10:58:53.091644 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:58:54 crc kubenswrapper[4733]: I0318 10:58:54.175587 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:55 crc kubenswrapper[4733]: I0318 10:58:55.115132 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373"} Mar 18 10:58:55 crc kubenswrapper[4733]: I0318 10:58:55.115707 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 10:58:57 crc kubenswrapper[4733]: I0318 10:58:57.563899 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:57 crc kubenswrapper[4733]: I0318 10:58:57.564311 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:57 crc kubenswrapper[4733]: I0318 10:58:57.670470 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:58 crc kubenswrapper[4733]: I0318 10:58:58.223333 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:58:58 crc kubenswrapper[4733]: I0318 10:58:58.295830 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:58:59 crc kubenswrapper[4733]: I0318 10:58:59.168746 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" exitCode=0 Mar 18 10:58:59 crc kubenswrapper[4733]: I0318 10:58:59.169025 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373"} Mar 18 10:58:59 crc kubenswrapper[4733]: I0318 10:58:59.169447 4733 scope.go:117] "RemoveContainer" containerID="fb87c31929d690ace0713c6e835580b64a2fb69bcf3837bfb62aeeeefbe16b5c" Mar 18 10:58:59 crc kubenswrapper[4733]: I0318 10:58:59.170299 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 10:58:59 crc kubenswrapper[4733]: E0318 10:58:59.170658 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.185897 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dpxzp" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="registry-server" containerID="cri-o://44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16" gracePeriod=2 Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.622600 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.694477 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content\") pod \"d2b9d483-b743-44ab-bf0b-f0b22f133576\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.694577 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4v5l\" (UniqueName: \"kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l\") pod \"d2b9d483-b743-44ab-bf0b-f0b22f133576\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.694656 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities\") pod \"d2b9d483-b743-44ab-bf0b-f0b22f133576\" (UID: \"d2b9d483-b743-44ab-bf0b-f0b22f133576\") " Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.695851 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities" (OuterVolumeSpecName: "utilities") pod "d2b9d483-b743-44ab-bf0b-f0b22f133576" (UID: "d2b9d483-b743-44ab-bf0b-f0b22f133576"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.704118 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l" (OuterVolumeSpecName: "kube-api-access-q4v5l") pod "d2b9d483-b743-44ab-bf0b-f0b22f133576" (UID: "d2b9d483-b743-44ab-bf0b-f0b22f133576"). InnerVolumeSpecName "kube-api-access-q4v5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.757139 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2b9d483-b743-44ab-bf0b-f0b22f133576" (UID: "d2b9d483-b743-44ab-bf0b-f0b22f133576"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.797216 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.797244 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4v5l\" (UniqueName: \"kubernetes.io/projected/d2b9d483-b743-44ab-bf0b-f0b22f133576-kube-api-access-q4v5l\") on node \"crc\" DevicePath \"\"" Mar 18 10:59:00 crc kubenswrapper[4733]: I0318 10:59:00.797258 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2b9d483-b743-44ab-bf0b-f0b22f133576-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.203745 4733 generic.go:334] "Generic (PLEG): container finished" podID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerID="44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16" exitCode=0 Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.203860 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerDied","Data":"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16"} Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.204289 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpxzp" event={"ID":"d2b9d483-b743-44ab-bf0b-f0b22f133576","Type":"ContainerDied","Data":"5d45d05271d7475d8f987b38020b7a838549e3626fa27dfe3061a1df224ec8fe"} Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.204331 4733 scope.go:117] "RemoveContainer" containerID="44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.203981 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpxzp" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.237723 4733 scope.go:117] "RemoveContainer" containerID="00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.263116 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.273329 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dpxzp"] Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.282764 4733 scope.go:117] "RemoveContainer" containerID="d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.327415 4733 scope.go:117] "RemoveContainer" containerID="44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16" Mar 18 10:59:01 crc kubenswrapper[4733]: E0318 10:59:01.327937 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16\": container with ID starting with 44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16 not found: ID does not exist" containerID="44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.327991 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16"} err="failed to get container status \"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16\": rpc error: code = NotFound desc = could not find container \"44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16\": container with ID starting with 44b4592be70be18f7abf508277d2b60867509d702745fc39e315eb1db7399b16 not found: ID does not exist" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.328026 4733 scope.go:117] "RemoveContainer" containerID="00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9" Mar 18 10:59:01 crc kubenswrapper[4733]: E0318 10:59:01.328559 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9\": container with ID starting with 00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9 not found: ID does not exist" containerID="00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.328600 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9"} err="failed to get container status \"00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9\": rpc error: code = NotFound desc = could not find container \"00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9\": container with ID starting with 00990b4a98758f8d96d9c56f65f35d6f4152d627814babba7b1c188d594f8df9 not found: ID does not exist" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.328624 4733 scope.go:117] "RemoveContainer" containerID="d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680" Mar 18 10:59:01 crc kubenswrapper[4733]: E0318 10:59:01.329029 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680\": container with ID starting with d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680 not found: ID does not exist" containerID="d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680" Mar 18 10:59:01 crc kubenswrapper[4733]: I0318 10:59:01.329100 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680"} err="failed to get container status \"d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680\": rpc error: code = NotFound desc = could not find container \"d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680\": container with ID starting with d16fb3f865634c313d04cd15fe257f949ffe967e94771f738a23b25fe3ead680 not found: ID does not exist" Mar 18 10:59:03 crc kubenswrapper[4733]: I0318 10:59:03.194821 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" path="/var/lib/kubelet/pods/d2b9d483-b743-44ab-bf0b-f0b22f133576/volumes" Mar 18 10:59:05 crc kubenswrapper[4733]: I0318 10:59:05.176151 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:59:05 crc kubenswrapper[4733]: E0318 10:59:05.176653 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:59:11 crc kubenswrapper[4733]: I0318 10:59:11.184602 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 10:59:11 crc kubenswrapper[4733]: E0318 10:59:11.185832 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:59:13 crc kubenswrapper[4733]: I0318 10:59:13.571737 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:59:13 crc kubenswrapper[4733]: I0318 10:59:13.572148 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:59:16 crc kubenswrapper[4733]: I0318 10:59:16.176098 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:59:16 crc kubenswrapper[4733]: E0318 10:59:16.177041 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:59:22 crc kubenswrapper[4733]: I0318 10:59:22.176163 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 10:59:22 crc kubenswrapper[4733]: E0318 10:59:22.177531 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:59:29 crc kubenswrapper[4733]: I0318 10:59:29.176170 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:59:29 crc kubenswrapper[4733]: E0318 10:59:29.177099 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:59:37 crc kubenswrapper[4733]: I0318 10:59:37.175424 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 10:59:37 crc kubenswrapper[4733]: E0318 10:59:37.176346 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:59:42 crc kubenswrapper[4733]: I0318 10:59:42.175272 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:59:42 crc kubenswrapper[4733]: E0318 10:59:42.176114 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 10:59:43 crc kubenswrapper[4733]: I0318 10:59:43.570914 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 10:59:43 crc kubenswrapper[4733]: I0318 10:59:43.571006 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 10:59:43 crc kubenswrapper[4733]: I0318 10:59:43.571100 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 10:59:43 crc kubenswrapper[4733]: I0318 10:59:43.572142 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 10:59:43 crc kubenswrapper[4733]: I0318 10:59:43.572343 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56" gracePeriod=600 Mar 18 10:59:44 crc kubenswrapper[4733]: I0318 10:59:44.662852 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56" exitCode=0 Mar 18 10:59:44 crc kubenswrapper[4733]: I0318 10:59:44.662911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56"} Mar 18 10:59:44 crc kubenswrapper[4733]: I0318 10:59:44.663456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9"} Mar 18 10:59:44 crc kubenswrapper[4733]: I0318 10:59:44.663481 4733 scope.go:117] "RemoveContainer" containerID="13079617ef56fbdc98c390ba5bdaff3c5530411f54f691fbeb11894744ecac48" Mar 18 10:59:52 crc kubenswrapper[4733]: I0318 10:59:52.176753 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 10:59:52 crc kubenswrapper[4733]: E0318 10:59:52.177914 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 10:59:57 crc kubenswrapper[4733]: I0318 10:59:57.175388 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 10:59:57 crc kubenswrapper[4733]: E0318 10:59:57.177337 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.177828 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s"] Mar 18 11:00:00 crc kubenswrapper[4733]: E0318 11:00:00.179031 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="registry-server" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.179053 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="registry-server" Mar 18 11:00:00 crc kubenswrapper[4733]: E0318 11:00:00.179085 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="extract-utilities" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.179100 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="extract-utilities" Mar 18 11:00:00 crc kubenswrapper[4733]: E0318 11:00:00.179162 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="extract-content" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.179177 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="extract-content" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.179546 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b9d483-b743-44ab-bf0b-f0b22f133576" containerName="registry-server" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.180560 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.183861 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.189783 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563860-2k4dj"] Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.191664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.193699 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.194018 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.200451 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.201311 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.212695 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s"] Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.219347 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563860-2k4dj"] Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.271921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgjxh\" (UniqueName: \"kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.272045 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.272122 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.373655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.373730 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.373757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqq4v\" (UniqueName: \"kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v\") pod \"auto-csr-approver-29563860-2k4dj\" (UID: \"6e41c70a-6d8d-47a8-9caf-57f46a60f96a\") " pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.373789 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgjxh\" (UniqueName: \"kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.374470 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.390844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgjxh\" (UniqueName: \"kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.393652 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume\") pod \"collect-profiles-29563860-dfn6s\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.474928 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqq4v\" (UniqueName: \"kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v\") pod \"auto-csr-approver-29563860-2k4dj\" (UID: \"6e41c70a-6d8d-47a8-9caf-57f46a60f96a\") " pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.492654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqq4v\" (UniqueName: \"kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v\") pod \"auto-csr-approver-29563860-2k4dj\" (UID: \"6e41c70a-6d8d-47a8-9caf-57f46a60f96a\") " pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.510885 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.535534 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:00 crc kubenswrapper[4733]: I0318 11:00:00.985199 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s"] Mar 18 11:00:00 crc kubenswrapper[4733]: W0318 11:00:00.993111 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf991012_f463_47ee_83f1_98bd34adcf1e.slice/crio-064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d WatchSource:0}: Error finding container 064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d: Status 404 returned error can't find the container with id 064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d Mar 18 11:00:01 crc kubenswrapper[4733]: I0318 11:00:01.057273 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563860-2k4dj"] Mar 18 11:00:01 crc kubenswrapper[4733]: W0318 11:00:01.061065 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e41c70a_6d8d_47a8_9caf_57f46a60f96a.slice/crio-d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593 WatchSource:0}: Error finding container d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593: Status 404 returned error can't find the container with id d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593 Mar 18 11:00:01 crc kubenswrapper[4733]: I0318 11:00:01.843257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" event={"ID":"6e41c70a-6d8d-47a8-9caf-57f46a60f96a","Type":"ContainerStarted","Data":"d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593"} Mar 18 11:00:01 crc kubenswrapper[4733]: I0318 11:00:01.845388 4733 generic.go:334] "Generic (PLEG): container finished" podID="bf991012-f463-47ee-83f1-98bd34adcf1e" containerID="2799cecba73aefe0ec40d27bcbc9d2213abfeaa8dc27c148695183a3c55debbd" exitCode=0 Mar 18 11:00:01 crc kubenswrapper[4733]: I0318 11:00:01.845452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" event={"ID":"bf991012-f463-47ee-83f1-98bd34adcf1e","Type":"ContainerDied","Data":"2799cecba73aefe0ec40d27bcbc9d2213abfeaa8dc27c148695183a3c55debbd"} Mar 18 11:00:01 crc kubenswrapper[4733]: I0318 11:00:01.845534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" event={"ID":"bf991012-f463-47ee-83f1-98bd34adcf1e","Type":"ContainerStarted","Data":"064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d"} Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.170115 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.345549 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume\") pod \"bf991012-f463-47ee-83f1-98bd34adcf1e\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.345632 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgjxh\" (UniqueName: \"kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh\") pod \"bf991012-f463-47ee-83f1-98bd34adcf1e\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.345818 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume\") pod \"bf991012-f463-47ee-83f1-98bd34adcf1e\" (UID: \"bf991012-f463-47ee-83f1-98bd34adcf1e\") " Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.346665 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume" (OuterVolumeSpecName: "config-volume") pod "bf991012-f463-47ee-83f1-98bd34adcf1e" (UID: "bf991012-f463-47ee-83f1-98bd34adcf1e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.351392 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bf991012-f463-47ee-83f1-98bd34adcf1e" (UID: "bf991012-f463-47ee-83f1-98bd34adcf1e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.352386 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh" (OuterVolumeSpecName: "kube-api-access-sgjxh") pod "bf991012-f463-47ee-83f1-98bd34adcf1e" (UID: "bf991012-f463-47ee-83f1-98bd34adcf1e"). InnerVolumeSpecName "kube-api-access-sgjxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.447381 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf991012-f463-47ee-83f1-98bd34adcf1e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.447625 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf991012-f463-47ee-83f1-98bd34adcf1e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.447639 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgjxh\" (UniqueName: \"kubernetes.io/projected/bf991012-f463-47ee-83f1-98bd34adcf1e-kube-api-access-sgjxh\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.862379 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" event={"ID":"bf991012-f463-47ee-83f1-98bd34adcf1e","Type":"ContainerDied","Data":"064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d"} Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.862425 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="064af1ddf7399134cda4e62b057d6eb2b8148ac3892e6bd388e41a03a7f1c70d" Mar 18 11:00:03 crc kubenswrapper[4733]: I0318 11:00:03.862448 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563860-dfn6s" Mar 18 11:00:04 crc kubenswrapper[4733]: I0318 11:00:04.175506 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:00:04 crc kubenswrapper[4733]: E0318 11:00:04.175967 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:00:04 crc kubenswrapper[4733]: I0318 11:00:04.267913 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6"] Mar 18 11:00:04 crc kubenswrapper[4733]: I0318 11:00:04.279559 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563815-tsrs6"] Mar 18 11:00:05 crc kubenswrapper[4733]: I0318 11:00:05.186106 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d915f7d2-5b4d-4017-a839-b615a182fafb" path="/var/lib/kubelet/pods/d915f7d2-5b4d-4017-a839-b615a182fafb/volumes" Mar 18 11:00:10 crc kubenswrapper[4733]: I0318 11:00:10.175925 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:00:10 crc kubenswrapper[4733]: E0318 11:00:10.176735 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:00:11 crc kubenswrapper[4733]: I0318 11:00:11.372121 4733 scope.go:117] "RemoveContainer" containerID="6068780e861c95e2a5524c6995b5943bf2eb924f4e716f49bfa978772d8dc58d" Mar 18 11:00:18 crc kubenswrapper[4733]: I0318 11:00:18.890784 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:18 crc kubenswrapper[4733]: E0318 11:00:18.891752 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf991012-f463-47ee-83f1-98bd34adcf1e" containerName="collect-profiles" Mar 18 11:00:18 crc kubenswrapper[4733]: I0318 11:00:18.891770 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf991012-f463-47ee-83f1-98bd34adcf1e" containerName="collect-profiles" Mar 18 11:00:18 crc kubenswrapper[4733]: I0318 11:00:18.892021 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf991012-f463-47ee-83f1-98bd34adcf1e" containerName="collect-profiles" Mar 18 11:00:18 crc kubenswrapper[4733]: I0318 11:00:18.893567 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:18 crc kubenswrapper[4733]: I0318 11:00:18.914410 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.014057 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.014123 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfptj\" (UniqueName: \"kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.014467 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.115686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.115744 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfptj\" (UniqueName: \"kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.115818 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.116395 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.116414 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.141436 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfptj\" (UniqueName: \"kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj\") pod \"redhat-operators-r62h8\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.175580 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:00:19 crc kubenswrapper[4733]: E0318 11:00:19.176001 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.222053 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:19 crc kubenswrapper[4733]: I0318 11:00:19.683912 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:20 crc kubenswrapper[4733]: I0318 11:00:20.002108 4733 generic.go:334] "Generic (PLEG): container finished" podID="6e41c70a-6d8d-47a8-9caf-57f46a60f96a" containerID="ee8a5931d088bb90e3f8edd41217a30f581b7d88c4f982136e16b0f2b145d28c" exitCode=0 Mar 18 11:00:20 crc kubenswrapper[4733]: I0318 11:00:20.002432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" event={"ID":"6e41c70a-6d8d-47a8-9caf-57f46a60f96a","Type":"ContainerDied","Data":"ee8a5931d088bb90e3f8edd41217a30f581b7d88c4f982136e16b0f2b145d28c"} Mar 18 11:00:20 crc kubenswrapper[4733]: I0318 11:00:20.004126 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerID="f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317" exitCode=0 Mar 18 11:00:20 crc kubenswrapper[4733]: I0318 11:00:20.004154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerDied","Data":"f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317"} Mar 18 11:00:20 crc kubenswrapper[4733]: I0318 11:00:20.004168 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerStarted","Data":"04995b1b3f36eef24f7f078917a3d8bd2c408f0c02eb5dd2635fe0b5f1c102f9"} Mar 18 11:00:21 crc kubenswrapper[4733]: I0318 11:00:21.385584 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:21 crc kubenswrapper[4733]: I0318 11:00:21.458249 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqq4v\" (UniqueName: \"kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v\") pod \"6e41c70a-6d8d-47a8-9caf-57f46a60f96a\" (UID: \"6e41c70a-6d8d-47a8-9caf-57f46a60f96a\") " Mar 18 11:00:21 crc kubenswrapper[4733]: I0318 11:00:21.463388 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v" (OuterVolumeSpecName: "kube-api-access-jqq4v") pod "6e41c70a-6d8d-47a8-9caf-57f46a60f96a" (UID: "6e41c70a-6d8d-47a8-9caf-57f46a60f96a"). InnerVolumeSpecName "kube-api-access-jqq4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:00:21 crc kubenswrapper[4733]: I0318 11:00:21.559811 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqq4v\" (UniqueName: \"kubernetes.io/projected/6e41c70a-6d8d-47a8-9caf-57f46a60f96a-kube-api-access-jqq4v\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.020473 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerID="104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904" exitCode=0 Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.020825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerDied","Data":"104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904"} Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.023319 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" event={"ID":"6e41c70a-6d8d-47a8-9caf-57f46a60f96a","Type":"ContainerDied","Data":"d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593"} Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.023365 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d414ba428ab0652aa66cf66c5272a3ace20f3bd338bfc094c7e21c52ebfae593" Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.023435 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563860-2k4dj" Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.464638 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563854-8xm7m"] Mar 18 11:00:22 crc kubenswrapper[4733]: I0318 11:00:22.475779 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563854-8xm7m"] Mar 18 11:00:23 crc kubenswrapper[4733]: I0318 11:00:23.190923 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2194f8e-8219-4d20-9657-bab035e9ce0b" path="/var/lib/kubelet/pods/e2194f8e-8219-4d20-9657-bab035e9ce0b/volumes" Mar 18 11:00:24 crc kubenswrapper[4733]: I0318 11:00:24.054791 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerStarted","Data":"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8"} Mar 18 11:00:24 crc kubenswrapper[4733]: I0318 11:00:24.085924 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r62h8" podStartSLOduration=2.927193043 podStartE2EDuration="6.085905467s" podCreationTimestamp="2026-03-18 11:00:18 +0000 UTC" firstStartedPulling="2026-03-18 11:00:20.005540102 +0000 UTC m=+2859.497274427" lastFinishedPulling="2026-03-18 11:00:23.164252536 +0000 UTC m=+2862.655986851" observedRunningTime="2026-03-18 11:00:24.076162782 +0000 UTC m=+2863.567897187" watchObservedRunningTime="2026-03-18 11:00:24.085905467 +0000 UTC m=+2863.577639802" Mar 18 11:00:24 crc kubenswrapper[4733]: I0318 11:00:24.177596 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:00:24 crc kubenswrapper[4733]: E0318 11:00:24.177835 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:00:29 crc kubenswrapper[4733]: I0318 11:00:29.222674 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:29 crc kubenswrapper[4733]: I0318 11:00:29.223350 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:30 crc kubenswrapper[4733]: I0318 11:00:30.280691 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r62h8" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="registry-server" probeResult="failure" output=< Mar 18 11:00:30 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Mar 18 11:00:30 crc kubenswrapper[4733]: > Mar 18 11:00:33 crc kubenswrapper[4733]: I0318 11:00:33.175984 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:00:33 crc kubenswrapper[4733]: E0318 11:00:33.176747 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:00:37 crc kubenswrapper[4733]: I0318 11:00:37.175379 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:00:37 crc kubenswrapper[4733]: E0318 11:00:37.177570 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:00:39 crc kubenswrapper[4733]: I0318 11:00:39.293454 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:39 crc kubenswrapper[4733]: I0318 11:00:39.364282 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:39 crc kubenswrapper[4733]: I0318 11:00:39.557252 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.213066 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r62h8" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="registry-server" containerID="cri-o://6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8" gracePeriod=2 Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.801764 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.920803 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities\") pod \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.920853 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content\") pod \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.920900 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfptj\" (UniqueName: \"kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj\") pod \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\" (UID: \"c9b1aa26-b953-4a50-a4b6-0edc261c5036\") " Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.922522 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities" (OuterVolumeSpecName: "utilities") pod "c9b1aa26-b953-4a50-a4b6-0edc261c5036" (UID: "c9b1aa26-b953-4a50-a4b6-0edc261c5036"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:00:41 crc kubenswrapper[4733]: I0318 11:00:41.932685 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj" (OuterVolumeSpecName: "kube-api-access-qfptj") pod "c9b1aa26-b953-4a50-a4b6-0edc261c5036" (UID: "c9b1aa26-b953-4a50-a4b6-0edc261c5036"). InnerVolumeSpecName "kube-api-access-qfptj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.022705 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.022740 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfptj\" (UniqueName: \"kubernetes.io/projected/c9b1aa26-b953-4a50-a4b6-0edc261c5036-kube-api-access-qfptj\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.099099 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9b1aa26-b953-4a50-a4b6-0edc261c5036" (UID: "c9b1aa26-b953-4a50-a4b6-0edc261c5036"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.124479 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1aa26-b953-4a50-a4b6-0edc261c5036-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.226828 4733 generic.go:334] "Generic (PLEG): container finished" podID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerID="6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8" exitCode=0 Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.226877 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r62h8" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.226878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerDied","Data":"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8"} Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.227011 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r62h8" event={"ID":"c9b1aa26-b953-4a50-a4b6-0edc261c5036","Type":"ContainerDied","Data":"04995b1b3f36eef24f7f078917a3d8bd2c408f0c02eb5dd2635fe0b5f1c102f9"} Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.227045 4733 scope.go:117] "RemoveContainer" containerID="6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.261922 4733 scope.go:117] "RemoveContainer" containerID="104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.282526 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.289562 4733 scope.go:117] "RemoveContainer" containerID="f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.292815 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r62h8"] Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.337821 4733 scope.go:117] "RemoveContainer" containerID="6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8" Mar 18 11:00:42 crc kubenswrapper[4733]: E0318 11:00:42.338740 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8\": container with ID starting with 6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8 not found: ID does not exist" containerID="6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.338781 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8"} err="failed to get container status \"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8\": rpc error: code = NotFound desc = could not find container \"6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8\": container with ID starting with 6d6befd46f4eb77f197da2b1ba26e4921971717d431e08ee39f2e013f94e03b8 not found: ID does not exist" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.338805 4733 scope.go:117] "RemoveContainer" containerID="104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904" Mar 18 11:00:42 crc kubenswrapper[4733]: E0318 11:00:42.339205 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904\": container with ID starting with 104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904 not found: ID does not exist" containerID="104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.339308 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904"} err="failed to get container status \"104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904\": rpc error: code = NotFound desc = could not find container \"104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904\": container with ID starting with 104ec438bd7e1a0d9263648df2d07431c5cfac4181c73f8b04e13a665edfe904 not found: ID does not exist" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.339411 4733 scope.go:117] "RemoveContainer" containerID="f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317" Mar 18 11:00:42 crc kubenswrapper[4733]: E0318 11:00:42.339781 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317\": container with ID starting with f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317 not found: ID does not exist" containerID="f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317" Mar 18 11:00:42 crc kubenswrapper[4733]: I0318 11:00:42.339803 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317"} err="failed to get container status \"f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317\": rpc error: code = NotFound desc = could not find container \"f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317\": container with ID starting with f0919ae919a0dce1024c8f5a9092475badc3c61b34605b876a869ff90201c317 not found: ID does not exist" Mar 18 11:00:43 crc kubenswrapper[4733]: I0318 11:00:43.198046 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" path="/var/lib/kubelet/pods/c9b1aa26-b953-4a50-a4b6-0edc261c5036/volumes" Mar 18 11:00:48 crc kubenswrapper[4733]: I0318 11:00:48.175828 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:00:48 crc kubenswrapper[4733]: E0318 11:00:48.177847 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:00:49 crc kubenswrapper[4733]: I0318 11:00:49.175707 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:00:49 crc kubenswrapper[4733]: E0318 11:00:49.176773 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:00:59 crc kubenswrapper[4733]: I0318 11:00:59.176016 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:00:59 crc kubenswrapper[4733]: E0318 11:00:59.176889 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:01:02 crc kubenswrapper[4733]: I0318 11:01:02.175675 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:01:02 crc kubenswrapper[4733]: E0318 11:01:02.176139 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.985336 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:05 crc kubenswrapper[4733]: E0318 11:01:05.985890 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="extract-utilities" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.985902 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="extract-utilities" Mar 18 11:01:05 crc kubenswrapper[4733]: E0318 11:01:05.985918 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="registry-server" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.985924 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="registry-server" Mar 18 11:01:05 crc kubenswrapper[4733]: E0318 11:01:05.985934 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="extract-content" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.985942 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="extract-content" Mar 18 11:01:05 crc kubenswrapper[4733]: E0318 11:01:05.985961 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e41c70a-6d8d-47a8-9caf-57f46a60f96a" containerName="oc" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.985966 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e41c70a-6d8d-47a8-9caf-57f46a60f96a" containerName="oc" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.986119 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b1aa26-b953-4a50-a4b6-0edc261c5036" containerName="registry-server" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.986133 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e41c70a-6d8d-47a8-9caf-57f46a60f96a" containerName="oc" Mar 18 11:01:05 crc kubenswrapper[4733]: I0318 11:01:05.987212 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.000283 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.063839 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5vxg\" (UniqueName: \"kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.063907 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.063972 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.166041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5vxg\" (UniqueName: \"kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.166119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.166207 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.166867 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.166928 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.190252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5vxg\" (UniqueName: \"kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg\") pod \"redhat-marketplace-h4rcn\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.312260 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:06 crc kubenswrapper[4733]: I0318 11:01:06.764159 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:07 crc kubenswrapper[4733]: I0318 11:01:07.468033 4733 generic.go:334] "Generic (PLEG): container finished" podID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerID="08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906" exitCode=0 Mar 18 11:01:07 crc kubenswrapper[4733]: I0318 11:01:07.468459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerDied","Data":"08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906"} Mar 18 11:01:07 crc kubenswrapper[4733]: I0318 11:01:07.468517 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerStarted","Data":"4f2a66f9800ceda513c83b3bfc142b08aff0f8d480daf6fc4b3befe4af03ca24"} Mar 18 11:01:07 crc kubenswrapper[4733]: I0318 11:01:07.472246 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 11:01:08 crc kubenswrapper[4733]: I0318 11:01:08.481119 4733 generic.go:334] "Generic (PLEG): container finished" podID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerID="72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58" exitCode=0 Mar 18 11:01:08 crc kubenswrapper[4733]: I0318 11:01:08.481224 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerDied","Data":"72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58"} Mar 18 11:01:09 crc kubenswrapper[4733]: I0318 11:01:09.490348 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerStarted","Data":"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb"} Mar 18 11:01:09 crc kubenswrapper[4733]: I0318 11:01:09.524897 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4rcn" podStartSLOduration=2.912135835 podStartE2EDuration="4.524878218s" podCreationTimestamp="2026-03-18 11:01:05 +0000 UTC" firstStartedPulling="2026-03-18 11:01:07.471980297 +0000 UTC m=+2906.963714622" lastFinishedPulling="2026-03-18 11:01:09.08472269 +0000 UTC m=+2908.576457005" observedRunningTime="2026-03-18 11:01:09.517630644 +0000 UTC m=+2909.009364979" watchObservedRunningTime="2026-03-18 11:01:09.524878218 +0000 UTC m=+2909.016612533" Mar 18 11:01:11 crc kubenswrapper[4733]: I0318 11:01:11.446721 4733 scope.go:117] "RemoveContainer" containerID="0646c2eb1d4076069ba17429b100767c9ea92208b7525c26c0789773916b849f" Mar 18 11:01:13 crc kubenswrapper[4733]: I0318 11:01:13.186361 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:01:13 crc kubenswrapper[4733]: E0318 11:01:13.186956 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.176715 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:01:16 crc kubenswrapper[4733]: E0318 11:01:16.177500 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.313127 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.313236 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.356416 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.616393 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:16 crc kubenswrapper[4733]: I0318 11:01:16.671851 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:18 crc kubenswrapper[4733]: I0318 11:01:18.576948 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4rcn" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="registry-server" containerID="cri-o://96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb" gracePeriod=2 Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.132655 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.195829 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities\") pod \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.195945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5vxg\" (UniqueName: \"kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg\") pod \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.196031 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content\") pod \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\" (UID: \"ea07c343-7a05-49fc-b5d6-4cbeda6a5381\") " Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.196877 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities" (OuterVolumeSpecName: "utilities") pod "ea07c343-7a05-49fc-b5d6-4cbeda6a5381" (UID: "ea07c343-7a05-49fc-b5d6-4cbeda6a5381"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.205456 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg" (OuterVolumeSpecName: "kube-api-access-b5vxg") pod "ea07c343-7a05-49fc-b5d6-4cbeda6a5381" (UID: "ea07c343-7a05-49fc-b5d6-4cbeda6a5381"). InnerVolumeSpecName "kube-api-access-b5vxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.220682 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea07c343-7a05-49fc-b5d6-4cbeda6a5381" (UID: "ea07c343-7a05-49fc-b5d6-4cbeda6a5381"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.298672 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.298770 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5vxg\" (UniqueName: \"kubernetes.io/projected/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-kube-api-access-b5vxg\") on node \"crc\" DevicePath \"\"" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.298833 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea07c343-7a05-49fc-b5d6-4cbeda6a5381-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.592748 4733 generic.go:334] "Generic (PLEG): container finished" podID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerID="96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb" exitCode=0 Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.592804 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerDied","Data":"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb"} Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.592839 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4rcn" event={"ID":"ea07c343-7a05-49fc-b5d6-4cbeda6a5381","Type":"ContainerDied","Data":"4f2a66f9800ceda513c83b3bfc142b08aff0f8d480daf6fc4b3befe4af03ca24"} Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.592841 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4rcn" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.592864 4733 scope.go:117] "RemoveContainer" containerID="96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.628412 4733 scope.go:117] "RemoveContainer" containerID="72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.655176 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.670170 4733 scope.go:117] "RemoveContainer" containerID="08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.675149 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4rcn"] Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.710020 4733 scope.go:117] "RemoveContainer" containerID="96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb" Mar 18 11:01:19 crc kubenswrapper[4733]: E0318 11:01:19.710984 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb\": container with ID starting with 96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb not found: ID does not exist" containerID="96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.711039 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb"} err="failed to get container status \"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb\": rpc error: code = NotFound desc = could not find container \"96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb\": container with ID starting with 96095daa5aad339dd34e460bfe35bb63755effddfba9c220193ced729dacf8bb not found: ID does not exist" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.711072 4733 scope.go:117] "RemoveContainer" containerID="72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58" Mar 18 11:01:19 crc kubenswrapper[4733]: E0318 11:01:19.711866 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58\": container with ID starting with 72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58 not found: ID does not exist" containerID="72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.711918 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58"} err="failed to get container status \"72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58\": rpc error: code = NotFound desc = could not find container \"72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58\": container with ID starting with 72be83bbac810419f54dcde0d39f65ddb52ff733b3ac42abe2f4c4d0f624ae58 not found: ID does not exist" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.711948 4733 scope.go:117] "RemoveContainer" containerID="08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906" Mar 18 11:01:19 crc kubenswrapper[4733]: E0318 11:01:19.712325 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906\": container with ID starting with 08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906 not found: ID does not exist" containerID="08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906" Mar 18 11:01:19 crc kubenswrapper[4733]: I0318 11:01:19.712376 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906"} err="failed to get container status \"08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906\": rpc error: code = NotFound desc = could not find container \"08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906\": container with ID starting with 08a3de0b8d7156f387872d738c867d86fe0dccd04c3114c5181b8c0ec179b906 not found: ID does not exist" Mar 18 11:01:21 crc kubenswrapper[4733]: I0318 11:01:21.195930 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" path="/var/lib/kubelet/pods/ea07c343-7a05-49fc-b5d6-4cbeda6a5381/volumes" Mar 18 11:01:26 crc kubenswrapper[4733]: I0318 11:01:26.176143 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:01:26 crc kubenswrapper[4733]: E0318 11:01:26.176899 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:01:28 crc kubenswrapper[4733]: I0318 11:01:28.176137 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:01:28 crc kubenswrapper[4733]: E0318 11:01:28.177041 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:01:41 crc kubenswrapper[4733]: I0318 11:01:41.185579 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:01:41 crc kubenswrapper[4733]: E0318 11:01:41.186718 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:01:42 crc kubenswrapper[4733]: I0318 11:01:42.175902 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:01:42 crc kubenswrapper[4733]: E0318 11:01:42.176265 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.571034 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.571475 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.699951 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m754c/must-gather-dtbq6"] Mar 18 11:01:43 crc kubenswrapper[4733]: E0318 11:01:43.700469 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="extract-content" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.700498 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="extract-content" Mar 18 11:01:43 crc kubenswrapper[4733]: E0318 11:01:43.700524 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="extract-utilities" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.700540 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="extract-utilities" Mar 18 11:01:43 crc kubenswrapper[4733]: E0318 11:01:43.700557 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="registry-server" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.700568 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="registry-server" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.700842 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea07c343-7a05-49fc-b5d6-4cbeda6a5381" containerName="registry-server" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.702234 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.704604 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m754c"/"openshift-service-ca.crt" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.707387 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m754c"/"kube-root-ca.crt" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.720875 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m754c/must-gather-dtbq6"] Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.838523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.838730 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57p52\" (UniqueName: \"kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.939843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.940011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57p52\" (UniqueName: \"kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.940298 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:43 crc kubenswrapper[4733]: I0318 11:01:43.958171 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57p52\" (UniqueName: \"kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52\") pod \"must-gather-dtbq6\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:44 crc kubenswrapper[4733]: I0318 11:01:44.022643 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:01:44 crc kubenswrapper[4733]: I0318 11:01:44.461964 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m754c/must-gather-dtbq6"] Mar 18 11:01:44 crc kubenswrapper[4733]: I0318 11:01:44.833451 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/must-gather-dtbq6" event={"ID":"5542a33f-3466-419c-af8f-3391bcc3d241","Type":"ContainerStarted","Data":"20467b25239c2c45218415caa32f35ca4a630d25fd4fdaa101490d758d81bb69"} Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.728096 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m754c/crc-debug-xs7wl"] Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.729580 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.733710 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m754c"/"default-dockercfg-rlbjp" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.824791 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz9mx\" (UniqueName: \"kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.824891 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.898372 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/must-gather-dtbq6" event={"ID":"5542a33f-3466-419c-af8f-3391bcc3d241","Type":"ContainerStarted","Data":"1557b68a8a325a0ed8ba1781990482335f57ef77fd7aa971706340af195f6c7b"} Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.898421 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/must-gather-dtbq6" event={"ID":"5542a33f-3466-419c-af8f-3391bcc3d241","Type":"ContainerStarted","Data":"83201b6bc07225e6541cd95837440f81a991cdbcc76e2856eb4af65abc081fb9"} Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.922599 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m754c/must-gather-dtbq6" podStartSLOduration=2.625467189 podStartE2EDuration="8.922582733s" podCreationTimestamp="2026-03-18 11:01:43 +0000 UTC" firstStartedPulling="2026-03-18 11:01:44.477491428 +0000 UTC m=+2943.969225793" lastFinishedPulling="2026-03-18 11:01:50.774607012 +0000 UTC m=+2950.266341337" observedRunningTime="2026-03-18 11:01:51.915800252 +0000 UTC m=+2951.407534587" watchObservedRunningTime="2026-03-18 11:01:51.922582733 +0000 UTC m=+2951.414317058" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.929104 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.929264 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz9mx\" (UniqueName: \"kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.929282 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:51 crc kubenswrapper[4733]: I0318 11:01:51.981505 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz9mx\" (UniqueName: \"kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx\") pod \"crc-debug-xs7wl\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:52 crc kubenswrapper[4733]: I0318 11:01:52.051172 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:01:52 crc kubenswrapper[4733]: I0318 11:01:52.910090 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/crc-debug-xs7wl" event={"ID":"cd690373-efdd-4c07-89b0-d7283af1e3eb","Type":"ContainerStarted","Data":"c106f3e82925e7c333ae1706bfb1aa321da818a6e980a043c0f60386d545ad29"} Mar 18 11:01:53 crc kubenswrapper[4733]: I0318 11:01:53.183153 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:01:53 crc kubenswrapper[4733]: E0318 11:01:53.183390 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:01:54 crc kubenswrapper[4733]: I0318 11:01:54.175474 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:01:54 crc kubenswrapper[4733]: E0318 11:01:54.176062 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.139717 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563862-vmt2n"] Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.140705 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.143563 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.143709 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.143804 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.151088 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563862-vmt2n"] Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.204328 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtn67\" (UniqueName: \"kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67\") pod \"auto-csr-approver-29563862-vmt2n\" (UID: \"ad06f0e3-f671-426a-b6d8-793e87745364\") " pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.305865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtn67\" (UniqueName: \"kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67\") pod \"auto-csr-approver-29563862-vmt2n\" (UID: \"ad06f0e3-f671-426a-b6d8-793e87745364\") " pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.335057 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtn67\" (UniqueName: \"kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67\") pod \"auto-csr-approver-29563862-vmt2n\" (UID: \"ad06f0e3-f671-426a-b6d8-793e87745364\") " pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:00 crc kubenswrapper[4733]: I0318 11:02:00.464095 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:06 crc kubenswrapper[4733]: W0318 11:02:06.155866 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad06f0e3_f671_426a_b6d8_793e87745364.slice/crio-217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7 WatchSource:0}: Error finding container 217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7: Status 404 returned error can't find the container with id 217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7 Mar 18 11:02:06 crc kubenswrapper[4733]: I0318 11:02:06.156393 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563862-vmt2n"] Mar 18 11:02:07 crc kubenswrapper[4733]: I0318 11:02:07.053331 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/crc-debug-xs7wl" event={"ID":"cd690373-efdd-4c07-89b0-d7283af1e3eb","Type":"ContainerStarted","Data":"d5fbd81f15e46718684cb2d4937c90c452a18fd504ebfaa9932047a8f5e25c6a"} Mar 18 11:02:07 crc kubenswrapper[4733]: I0318 11:02:07.055054 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" event={"ID":"ad06f0e3-f671-426a-b6d8-793e87745364","Type":"ContainerStarted","Data":"217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7"} Mar 18 11:02:07 crc kubenswrapper[4733]: I0318 11:02:07.178886 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:02:07 crc kubenswrapper[4733]: E0318 11:02:07.179342 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:02:07 crc kubenswrapper[4733]: I0318 11:02:07.179473 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:02:07 crc kubenswrapper[4733]: E0318 11:02:07.179865 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:02:08 crc kubenswrapper[4733]: I0318 11:02:08.064752 4733 generic.go:334] "Generic (PLEG): container finished" podID="ad06f0e3-f671-426a-b6d8-793e87745364" containerID="8d69124510c80bc1e6643920b76340f2795eccc769fbdff3665f6edcc2793fe5" exitCode=0 Mar 18 11:02:08 crc kubenswrapper[4733]: I0318 11:02:08.064822 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" event={"ID":"ad06f0e3-f671-426a-b6d8-793e87745364","Type":"ContainerDied","Data":"8d69124510c80bc1e6643920b76340f2795eccc769fbdff3665f6edcc2793fe5"} Mar 18 11:02:08 crc kubenswrapper[4733]: I0318 11:02:08.088598 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m754c/crc-debug-xs7wl" podStartSLOduration=3.2594349080000002 podStartE2EDuration="17.088572607s" podCreationTimestamp="2026-03-18 11:01:51 +0000 UTC" firstStartedPulling="2026-03-18 11:01:52.092250426 +0000 UTC m=+2951.583984751" lastFinishedPulling="2026-03-18 11:02:05.921388125 +0000 UTC m=+2965.413122450" observedRunningTime="2026-03-18 11:02:07.073998168 +0000 UTC m=+2966.565732513" watchObservedRunningTime="2026-03-18 11:02:08.088572607 +0000 UTC m=+2967.580306972" Mar 18 11:02:09 crc kubenswrapper[4733]: I0318 11:02:09.406194 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:09 crc kubenswrapper[4733]: I0318 11:02:09.553422 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtn67\" (UniqueName: \"kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67\") pod \"ad06f0e3-f671-426a-b6d8-793e87745364\" (UID: \"ad06f0e3-f671-426a-b6d8-793e87745364\") " Mar 18 11:02:09 crc kubenswrapper[4733]: I0318 11:02:09.558791 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67" (OuterVolumeSpecName: "kube-api-access-rtn67") pod "ad06f0e3-f671-426a-b6d8-793e87745364" (UID: "ad06f0e3-f671-426a-b6d8-793e87745364"). InnerVolumeSpecName "kube-api-access-rtn67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:02:09 crc kubenswrapper[4733]: I0318 11:02:09.655666 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtn67\" (UniqueName: \"kubernetes.io/projected/ad06f0e3-f671-426a-b6d8-793e87745364-kube-api-access-rtn67\") on node \"crc\" DevicePath \"\"" Mar 18 11:02:10 crc kubenswrapper[4733]: I0318 11:02:10.081559 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" event={"ID":"ad06f0e3-f671-426a-b6d8-793e87745364","Type":"ContainerDied","Data":"217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7"} Mar 18 11:02:10 crc kubenswrapper[4733]: I0318 11:02:10.081604 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="217b84ae69a58013cd13fbbc295a76316974fb46cf597690074f4cafa94bc5c7" Mar 18 11:02:10 crc kubenswrapper[4733]: I0318 11:02:10.081609 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563862-vmt2n" Mar 18 11:02:10 crc kubenswrapper[4733]: I0318 11:02:10.503927 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563856-qhf5l"] Mar 18 11:02:10 crc kubenswrapper[4733]: I0318 11:02:10.510403 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563856-qhf5l"] Mar 18 11:02:11 crc kubenswrapper[4733]: I0318 11:02:11.192265 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1a4900e-15be-4e3f-a8a2-3eb582acbc20" path="/var/lib/kubelet/pods/e1a4900e-15be-4e3f-a8a2-3eb582acbc20/volumes" Mar 18 11:02:11 crc kubenswrapper[4733]: I0318 11:02:11.544457 4733 scope.go:117] "RemoveContainer" containerID="9c09a3cb9db31583aa867b9b2e7873c25af33c3cb06cde66bc60959b1e039850" Mar 18 11:02:13 crc kubenswrapper[4733]: I0318 11:02:13.570954 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 11:02:13 crc kubenswrapper[4733]: I0318 11:02:13.571258 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 11:02:20 crc kubenswrapper[4733]: I0318 11:02:20.176096 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:02:20 crc kubenswrapper[4733]: E0318 11:02:20.177253 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:02:21 crc kubenswrapper[4733]: I0318 11:02:21.190140 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:02:21 crc kubenswrapper[4733]: E0318 11:02:21.191037 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:02:25 crc kubenswrapper[4733]: I0318 11:02:25.632872 4733 generic.go:334] "Generic (PLEG): container finished" podID="cd690373-efdd-4c07-89b0-d7283af1e3eb" containerID="d5fbd81f15e46718684cb2d4937c90c452a18fd504ebfaa9932047a8f5e25c6a" exitCode=0 Mar 18 11:02:25 crc kubenswrapper[4733]: I0318 11:02:25.632965 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/crc-debug-xs7wl" event={"ID":"cd690373-efdd-4c07-89b0-d7283af1e3eb","Type":"ContainerDied","Data":"d5fbd81f15e46718684cb2d4937c90c452a18fd504ebfaa9932047a8f5e25c6a"} Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.726381 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.760239 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m754c/crc-debug-xs7wl"] Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.765982 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m754c/crc-debug-xs7wl"] Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.846250 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host\") pod \"cd690373-efdd-4c07-89b0-d7283af1e3eb\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.846565 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz9mx\" (UniqueName: \"kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx\") pod \"cd690373-efdd-4c07-89b0-d7283af1e3eb\" (UID: \"cd690373-efdd-4c07-89b0-d7283af1e3eb\") " Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.846446 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host" (OuterVolumeSpecName: "host") pod "cd690373-efdd-4c07-89b0-d7283af1e3eb" (UID: "cd690373-efdd-4c07-89b0-d7283af1e3eb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.846860 4733 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd690373-efdd-4c07-89b0-d7283af1e3eb-host\") on node \"crc\" DevicePath \"\"" Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.856377 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx" (OuterVolumeSpecName: "kube-api-access-fz9mx") pod "cd690373-efdd-4c07-89b0-d7283af1e3eb" (UID: "cd690373-efdd-4c07-89b0-d7283af1e3eb"). InnerVolumeSpecName "kube-api-access-fz9mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:02:26 crc kubenswrapper[4733]: I0318 11:02:26.948475 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz9mx\" (UniqueName: \"kubernetes.io/projected/cd690373-efdd-4c07-89b0-d7283af1e3eb-kube-api-access-fz9mx\") on node \"crc\" DevicePath \"\"" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.185930 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd690373-efdd-4c07-89b0-d7283af1e3eb" path="/var/lib/kubelet/pods/cd690373-efdd-4c07-89b0-d7283af1e3eb/volumes" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.649453 4733 scope.go:117] "RemoveContainer" containerID="d5fbd81f15e46718684cb2d4937c90c452a18fd504ebfaa9932047a8f5e25c6a" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.649500 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-xs7wl" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.952821 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m754c/crc-debug-8v78b"] Mar 18 11:02:27 crc kubenswrapper[4733]: E0318 11:02:27.953216 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd690373-efdd-4c07-89b0-d7283af1e3eb" containerName="container-00" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.953233 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd690373-efdd-4c07-89b0-d7283af1e3eb" containerName="container-00" Mar 18 11:02:27 crc kubenswrapper[4733]: E0318 11:02:27.953278 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad06f0e3-f671-426a-b6d8-793e87745364" containerName="oc" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.953286 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad06f0e3-f671-426a-b6d8-793e87745364" containerName="oc" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.953459 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad06f0e3-f671-426a-b6d8-793e87745364" containerName="oc" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.953473 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd690373-efdd-4c07-89b0-d7283af1e3eb" containerName="container-00" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.954056 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.956656 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m754c"/"default-dockercfg-rlbjp" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.964034 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gstb\" (UniqueName: \"kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:27 crc kubenswrapper[4733]: I0318 11:02:27.964107 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.065682 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gstb\" (UniqueName: \"kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.066004 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.066120 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.090188 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gstb\" (UniqueName: \"kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb\") pod \"crc-debug-8v78b\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.278421 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.658695 4733 generic.go:334] "Generic (PLEG): container finished" podID="60216bd7-dbfa-4ef9-a60d-0941dd33ff44" containerID="de436dca9aa9677c82c7c01e372254e32bace5a1643be11aa46c20d0852dd35c" exitCode=1 Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.658807 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/crc-debug-8v78b" event={"ID":"60216bd7-dbfa-4ef9-a60d-0941dd33ff44","Type":"ContainerDied","Data":"de436dca9aa9677c82c7c01e372254e32bace5a1643be11aa46c20d0852dd35c"} Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.659021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/crc-debug-8v78b" event={"ID":"60216bd7-dbfa-4ef9-a60d-0941dd33ff44","Type":"ContainerStarted","Data":"696ddf380a049e8177e4e3e6b95841c6aafcaa12c024e09fdf4c74f937470a0e"} Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.692211 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m754c/crc-debug-8v78b"] Mar 18 11:02:28 crc kubenswrapper[4733]: I0318 11:02:28.703164 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m754c/crc-debug-8v78b"] Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.747534 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.896112 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gstb\" (UniqueName: \"kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb\") pod \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.896537 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host\") pod \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\" (UID: \"60216bd7-dbfa-4ef9-a60d-0941dd33ff44\") " Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.896663 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host" (OuterVolumeSpecName: "host") pod "60216bd7-dbfa-4ef9-a60d-0941dd33ff44" (UID: "60216bd7-dbfa-4ef9-a60d-0941dd33ff44"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.897270 4733 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-host\") on node \"crc\" DevicePath \"\"" Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.901888 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb" (OuterVolumeSpecName: "kube-api-access-6gstb") pod "60216bd7-dbfa-4ef9-a60d-0941dd33ff44" (UID: "60216bd7-dbfa-4ef9-a60d-0941dd33ff44"). InnerVolumeSpecName "kube-api-access-6gstb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:02:29 crc kubenswrapper[4733]: I0318 11:02:29.998254 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gstb\" (UniqueName: \"kubernetes.io/projected/60216bd7-dbfa-4ef9-a60d-0941dd33ff44-kube-api-access-6gstb\") on node \"crc\" DevicePath \"\"" Mar 18 11:02:30 crc kubenswrapper[4733]: I0318 11:02:30.677206 4733 scope.go:117] "RemoveContainer" containerID="de436dca9aa9677c82c7c01e372254e32bace5a1643be11aa46c20d0852dd35c" Mar 18 11:02:30 crc kubenswrapper[4733]: I0318 11:02:30.677969 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/crc-debug-8v78b" Mar 18 11:02:31 crc kubenswrapper[4733]: I0318 11:02:31.222461 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60216bd7-dbfa-4ef9-a60d-0941dd33ff44" path="/var/lib/kubelet/pods/60216bd7-dbfa-4ef9-a60d-0941dd33ff44/volumes" Mar 18 11:02:35 crc kubenswrapper[4733]: I0318 11:02:35.176120 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:02:35 crc kubenswrapper[4733]: E0318 11:02:35.178314 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:02:36 crc kubenswrapper[4733]: I0318 11:02:36.175625 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:02:36 crc kubenswrapper[4733]: E0318 11:02:36.176669 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.571105 4733 patch_prober.go:28] interesting pod/machine-config-daemon-2h7dp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.571769 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.571833 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.572759 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9"} pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.572848 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" containerName="machine-config-daemon" containerID="cri-o://6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" gracePeriod=600 Mar 18 11:02:43 crc kubenswrapper[4733]: E0318 11:02:43.722698 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.777919 4733 generic.go:334] "Generic (PLEG): container finished" podID="6f75e1c5-e0c5-43df-944f-77b734070793" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" exitCode=0 Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.777977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerDied","Data":"6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9"} Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.778011 4733 scope.go:117] "RemoveContainer" containerID="32198f7b4110f4b23718a4e872dd512bdbf76e8166cae4cab128ee6761e36a56" Mar 18 11:02:43 crc kubenswrapper[4733]: I0318 11:02:43.778576 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:02:43 crc kubenswrapper[4733]: E0318 11:02:43.778893 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.232713 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5f59b8f679-z8m4g_5fcd9264-61af-4872-82e6-8b0e1667ac70/init/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.436184 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5f59b8f679-z8m4g_5fcd9264-61af-4872-82e6-8b0e1667ac70/dnsmasq-dns/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.486035 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5f59b8f679-z8m4g_5fcd9264-61af-4872-82e6-8b0e1667ac70/init/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.541565 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_55f0b230-09f2-4be2-aa1f-76a37f3fe30c/kube-state-metrics/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.724236 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dd66892e-808c-405a-ac8e-366b6ca8b148/memcached/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.741151 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0208d826-df0f-41c8-83a7-821a21b7b85d/mysql-bootstrap/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.916137 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0208d826-df0f-41c8-83a7-821a21b7b85d/galera/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.929752 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_dc60b49b-96fa-40fd-a8e5-40c810f5ef80/mysql-bootstrap/0.log" Mar 18 11:02:46 crc kubenswrapper[4733]: I0318 11:02:46.975463 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0208d826-df0f-41c8-83a7-821a21b7b85d/mysql-bootstrap/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.093667 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_dc60b49b-96fa-40fd-a8e5-40c810f5ef80/mysql-bootstrap/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.098408 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_dc60b49b-96fa-40fd-a8e5-40c810f5ef80/galera/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.145872 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6trms_e7849feb-5f1b-4b67-a3f7-8a419ebda0bd/openstack-network-exporter/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.283605 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ljrgt_d75a8d54-aca8-49cd-9062-6389baaf7a09/ovsdb-server-init/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.442577 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ljrgt_d75a8d54-aca8-49cd-9062-6389baaf7a09/ovsdb-server-init/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.453835 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ljrgt_d75a8d54-aca8-49cd-9062-6389baaf7a09/ovs-vswitchd/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.468138 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ljrgt_d75a8d54-aca8-49cd-9062-6389baaf7a09/ovsdb-server/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.640251 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rh64b_e3c842d3-b3dd-4cf2-9df0-16cea4061bc5/ovn-controller/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.641094 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_96c7007d-b722-4518-a298-269808d7dfc5/openstack-network-exporter/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.673834 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_96c7007d-b722-4518-a298-269808d7dfc5/ovn-northd/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.796710 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a8c27598-870d-4de0-a986-47042d7d6f4c/openstack-network-exporter/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.823069 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a8c27598-870d-4de0-a986-47042d7d6f4c/ovsdbserver-nb/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.961767 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0868210e-9d93-4f63-b425-7db21f13cd90/openstack-network-exporter/0.log" Mar 18 11:02:47 crc kubenswrapper[4733]: I0318 11:02:47.972573 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0868210e-9d93-4f63-b425-7db21f13cd90/ovsdbserver-sb/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.059624 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4a4e3e2-bd4d-4f8d-97bc-51267378ab03/setup-container/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.179902 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4a4e3e2-bd4d-4f8d-97bc-51267378ab03/setup-container/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.214657 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4a4e3e2-bd4d-4f8d-97bc-51267378ab03/rabbitmq/10.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.269708 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b4a4e3e2-bd4d-4f8d-97bc-51267378ab03/rabbitmq/10.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.357135 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f0570ce4-1455-4698-85cf-01f7108d9e7f/setup-container/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.525039 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f0570ce4-1455-4698-85cf-01f7108d9e7f/setup-container/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.530629 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f0570ce4-1455-4698-85cf-01f7108d9e7f/rabbitmq/10.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.537091 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f0570ce4-1455-4698-85cf-01f7108d9e7f/rabbitmq/10.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.692878 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nfmp2_5e3fc960-7783-4952-90c9-1551c780ae03/swift-ring-rebalance/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.767017 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/account-auditor/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.785864 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/account-reaper/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.892101 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/account-replicator/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.893572 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/account-server/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.931159 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/container-auditor/0.log" Mar 18 11:02:48 crc kubenswrapper[4733]: I0318 11:02:48.992499 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/container-replicator/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.047893 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/container-updater/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.076537 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/container-server/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.106908 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/object-auditor/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.176168 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:02:49 crc kubenswrapper[4733]: E0318 11:02:49.176502 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.177078 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:02:49 crc kubenswrapper[4733]: E0318 11:02:49.177291 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.189921 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/object-expirer/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.207178 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/object-replicator/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.279627 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/object-server/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.306608 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/object-updater/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.388504 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/rsync/0.log" Mar 18 11:02:49 crc kubenswrapper[4733]: I0318 11:02:49.410385 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4f94cfc9-67cc-474c-8d99-58a9d4e0273f/swift-recon-cron/0.log" Mar 18 11:02:59 crc kubenswrapper[4733]: I0318 11:02:59.175792 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:02:59 crc kubenswrapper[4733]: E0318 11:02:59.176550 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:03:02 crc kubenswrapper[4733]: I0318 11:03:02.175802 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:03:02 crc kubenswrapper[4733]: E0318 11:03:02.176333 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:03:02 crc kubenswrapper[4733]: I0318 11:03:02.176366 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:03:02 crc kubenswrapper[4733]: E0318 11:03:02.176504 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:03:05 crc kubenswrapper[4733]: I0318 11:03:05.791831 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/util/0.log" Mar 18 11:03:05 crc kubenswrapper[4733]: I0318 11:03:05.968257 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/util/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.012148 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/pull/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.046648 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/pull/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.152158 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/util/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.168728 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/pull/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.186203 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_77b7da525d1abc8152b798823d798d773ced7ad76161af6957e3c157386hj67_53c111d7-ea42-4913-b378-ec44062b0691/extract/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.342484 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-sfv8v_0fb2ba68-fa0f-4483-afdf-2eb381c54320/manager/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.510605 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-t8796_748f4855-3978-4ecd-805e-0fee34ce0094/manager/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.720613 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-ljvrt_bc0e28fc-cff0-4c39-8073-61d5d6481866/manager/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.815641 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-cxlns_bd5ae902-d036-4e52-983d-aa3e1a86dca8/manager/0.log" Mar 18 11:03:06 crc kubenswrapper[4733]: I0318 11:03:06.957993 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-wkjtf_838f8a80-01c0-41d8-b431-2a23c9235fab/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.088894 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-v6zxn_8fe910c4-798b-4381-a71d-697459f7f79a/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.221749 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-74c694b97b-j4snz_651c7dd5-3adc-48b4-b579-309258aa3735/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.253509 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-pcscc_fd146b1e-59a9-4246-9520-f2d6f6cf6cd1/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.446674 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-chmbd_ae8a8fbc-d425-4da5-afb3-438a85a43722/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.459667 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-tp4s7_de7565f5-677b-4aeb-90ab-0d632b28b295/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.616026 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-4xzlc_79dfdcde-0538-4777-959e-1daf2b6263de/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.653585 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-gkndg_216f9239-7d2e-483e-a89f-0955a518aa4a/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.816457 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-jmwdk_f93025ae-ebc3-4aed-bfde-e514d8b814ce/manager/0.log" Mar 18 11:03:07 crc kubenswrapper[4733]: I0318 11:03:07.858485 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-22wt5_31999dbe-554e-4168-a902-1f62e82ce854/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.011410 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-s6rbv_6eca2f16-53b8-4173-ace4-18b7292b1369/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.142842 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-579f7bfb88-sfsb4_d1b10458-2335-4b46-9f63-c8a005096ff7/operator/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.338955 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-g2m9r_45605961-e7c2-4bd3-a670-d8541124408a/registry-server/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.495275 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-85877db48-qvlf2_a4b7e706-a9a7-490a-84a8-094d1d909ba8/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.512988 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-flv24_6762c515-b422-4157-a8ce-b9ca4781e134/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.652066 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-9txbj_4ad2d88a-c733-4409-b07b-5ff4661e1b68/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.705088 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-k64ch_e64c7cd6-a04b-440e-ac47-40f672fbc333/operator/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.858669 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-v2pb2_6ea742ac-3be9-4067-ab5a-032365494fde/manager/0.log" Mar 18 11:03:08 crc kubenswrapper[4733]: I0318 11:03:08.953874 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-fd4t7_759f85a1-4e24-4b61-879b-90801d648683/manager/0.log" Mar 18 11:03:09 crc kubenswrapper[4733]: I0318 11:03:09.031588 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-nskpj_6152e0d7-6362-4c7d-ba2b-4a1e55ca4f54/manager/0.log" Mar 18 11:03:09 crc kubenswrapper[4733]: I0318 11:03:09.145462 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-sqr4g_cd9234ed-fcbc-4d81-9034-27d39b3df6ee/manager/0.log" Mar 18 11:03:11 crc kubenswrapper[4733]: I0318 11:03:11.180523 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:03:11 crc kubenswrapper[4733]: E0318 11:03:11.181157 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:03:15 crc kubenswrapper[4733]: I0318 11:03:15.176132 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:03:15 crc kubenswrapper[4733]: E0318 11:03:15.176694 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:03:17 crc kubenswrapper[4733]: I0318 11:03:17.180915 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:03:17 crc kubenswrapper[4733]: E0318 11:03:17.181125 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:03:26 crc kubenswrapper[4733]: I0318 11:03:26.176964 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:03:26 crc kubenswrapper[4733]: E0318 11:03:26.178292 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:03:28 crc kubenswrapper[4733]: I0318 11:03:28.176267 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:03:28 crc kubenswrapper[4733]: E0318 11:03:28.176694 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:03:28 crc kubenswrapper[4733]: I0318 11:03:28.177239 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:03:28 crc kubenswrapper[4733]: E0318 11:03:28.177422 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:03:28 crc kubenswrapper[4733]: I0318 11:03:28.458622 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4lbr5_f2b6c2ec-c07f-4d59-ba90-1ed2ec55d8a7/control-plane-machine-set-operator/0.log" Mar 18 11:03:28 crc kubenswrapper[4733]: I0318 11:03:28.631944 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nbftd_0c02459c-3d75-4363-a010-3e9639bb9b4e/kube-rbac-proxy/0.log" Mar 18 11:03:28 crc kubenswrapper[4733]: I0318 11:03:28.647209 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nbftd_0c02459c-3d75-4363-a010-3e9639bb9b4e/machine-api-operator/0.log" Mar 18 11:03:37 crc kubenswrapper[4733]: I0318 11:03:37.175467 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:03:37 crc kubenswrapper[4733]: E0318 11:03:37.176689 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:03:39 crc kubenswrapper[4733]: I0318 11:03:39.179362 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:03:39 crc kubenswrapper[4733]: E0318 11:03:39.187617 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:03:40 crc kubenswrapper[4733]: I0318 11:03:40.176219 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:03:40 crc kubenswrapper[4733]: E0318 11:03:40.176403 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:03:40 crc kubenswrapper[4733]: I0318 11:03:40.964946 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-rd2dh_ce77d29d-b82e-46be-a694-b6eea5da9379/cert-manager-controller/0.log" Mar 18 11:03:41 crc kubenswrapper[4733]: I0318 11:03:41.155740 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-8ds68_585c06be-86bd-48b7-954e-9aec01b08874/cert-manager-cainjector/0.log" Mar 18 11:03:41 crc kubenswrapper[4733]: I0318 11:03:41.218884 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-2nr27_534b0ac6-c9b1-4940-9e6e-ed36de1ec1e8/cert-manager-webhook/0.log" Mar 18 11:03:51 crc kubenswrapper[4733]: I0318 11:03:51.181636 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:03:51 crc kubenswrapper[4733]: E0318 11:03:51.182283 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:03:52 crc kubenswrapper[4733]: I0318 11:03:52.176013 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:03:52 crc kubenswrapper[4733]: I0318 11:03:52.176506 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:03:52 crc kubenswrapper[4733]: E0318 11:03:52.176711 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:03:52 crc kubenswrapper[4733]: E0318 11:03:52.176939 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.480114 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-btpf9_95b678ac-c7be-4c57-8663-05b207f43338/nmstate-console-plugin/0.log" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.539826 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-8jncr_4c5d76ae-c917-4ba7-91d7-332a8e578245/nmstate-handler/0.log" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.589891 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-7swn6_eb2e5225-c943-4b06-b2de-90ab1168242b/kube-rbac-proxy/0.log" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.667749 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-7swn6_eb2e5225-c943-4b06-b2de-90ab1168242b/nmstate-metrics/0.log" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.771463 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-2d4dp_7c8f098b-42c0-4132-88c0-350e0c872f9d/nmstate-operator/0.log" Mar 18 11:03:53 crc kubenswrapper[4733]: I0318 11:03:53.853935 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-m6rhx_29460af7-7801-4268-aae8-f84763762e2f/nmstate-webhook/0.log" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.169393 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563864-g5m6l"] Mar 18 11:04:00 crc kubenswrapper[4733]: E0318 11:04:00.172090 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60216bd7-dbfa-4ef9-a60d-0941dd33ff44" containerName="container-00" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.176106 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="60216bd7-dbfa-4ef9-a60d-0941dd33ff44" containerName="container-00" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.176514 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="60216bd7-dbfa-4ef9-a60d-0941dd33ff44" containerName="container-00" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.177284 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.179621 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.179740 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.180596 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563864-g5m6l"] Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.180911 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.221789 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpdk7\" (UniqueName: \"kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7\") pod \"auto-csr-approver-29563864-g5m6l\" (UID: \"720459c9-dd9a-4d0f-8541-f4f2f578acc5\") " pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.324102 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpdk7\" (UniqueName: \"kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7\") pod \"auto-csr-approver-29563864-g5m6l\" (UID: \"720459c9-dd9a-4d0f-8541-f4f2f578acc5\") " pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.351740 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpdk7\" (UniqueName: \"kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7\") pod \"auto-csr-approver-29563864-g5m6l\" (UID: \"720459c9-dd9a-4d0f-8541-f4f2f578acc5\") " pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.498157 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:00 crc kubenswrapper[4733]: I0318 11:04:00.969496 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563864-g5m6l"] Mar 18 11:04:01 crc kubenswrapper[4733]: I0318 11:04:01.520755 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" event={"ID":"720459c9-dd9a-4d0f-8541-f4f2f578acc5","Type":"ContainerStarted","Data":"f347f95424cf6cf3dae5086a5ebf375241fc5353eb08c3a8685b22e00d50d8be"} Mar 18 11:04:03 crc kubenswrapper[4733]: I0318 11:04:03.535086 4733 generic.go:334] "Generic (PLEG): container finished" podID="720459c9-dd9a-4d0f-8541-f4f2f578acc5" containerID="b3d994fcf267bc98b12bd59da2b08ea67ce03ac437e382c7728b7a3d005bd1f9" exitCode=0 Mar 18 11:04:03 crc kubenswrapper[4733]: I0318 11:04:03.535142 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" event={"ID":"720459c9-dd9a-4d0f-8541-f4f2f578acc5","Type":"ContainerDied","Data":"b3d994fcf267bc98b12bd59da2b08ea67ce03ac437e382c7728b7a3d005bd1f9"} Mar 18 11:04:04 crc kubenswrapper[4733]: I0318 11:04:04.175754 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:04:04 crc kubenswrapper[4733]: I0318 11:04:04.544788 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01"} Mar 18 11:04:04 crc kubenswrapper[4733]: I0318 11:04:04.545446 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 11:04:04 crc kubenswrapper[4733]: I0318 11:04:04.867738 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.021821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpdk7\" (UniqueName: \"kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7\") pod \"720459c9-dd9a-4d0f-8541-f4f2f578acc5\" (UID: \"720459c9-dd9a-4d0f-8541-f4f2f578acc5\") " Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.026356 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7" (OuterVolumeSpecName: "kube-api-access-hpdk7") pod "720459c9-dd9a-4d0f-8541-f4f2f578acc5" (UID: "720459c9-dd9a-4d0f-8541-f4f2f578acc5"). InnerVolumeSpecName "kube-api-access-hpdk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.123811 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpdk7\" (UniqueName: \"kubernetes.io/projected/720459c9-dd9a-4d0f-8541-f4f2f578acc5-kube-api-access-hpdk7\") on node \"crc\" DevicePath \"\"" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.178747 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:04:05 crc kubenswrapper[4733]: E0318 11:04:05.179103 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.552825 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.552841 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563864-g5m6l" event={"ID":"720459c9-dd9a-4d0f-8541-f4f2f578acc5","Type":"ContainerDied","Data":"f347f95424cf6cf3dae5086a5ebf375241fc5353eb08c3a8685b22e00d50d8be"} Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.552885 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f347f95424cf6cf3dae5086a5ebf375241fc5353eb08c3a8685b22e00d50d8be" Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.946366 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563858-bxq6r"] Mar 18 11:04:05 crc kubenswrapper[4733]: I0318 11:04:05.952168 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563858-bxq6r"] Mar 18 11:04:07 crc kubenswrapper[4733]: I0318 11:04:07.176203 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:04:07 crc kubenswrapper[4733]: I0318 11:04:07.191689 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc61f2d-1837-4253-a3a3-91d8acc950f8" path="/var/lib/kubelet/pods/7bc61f2d-1837-4253-a3a3-91d8acc950f8/volumes" Mar 18 11:04:07 crc kubenswrapper[4733]: I0318 11:04:07.568561 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2"} Mar 18 11:04:07 crc kubenswrapper[4733]: I0318 11:04:07.569048 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 11:04:09 crc kubenswrapper[4733]: I0318 11:04:09.589615 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" exitCode=0 Mar 18 11:04:09 crc kubenswrapper[4733]: I0318 11:04:09.589656 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01"} Mar 18 11:04:09 crc kubenswrapper[4733]: I0318 11:04:09.589694 4733 scope.go:117] "RemoveContainer" containerID="42f5e854566e19360cf16fa02f3e09efbcbafeba0ef62811eb321face5cd1f9f" Mar 18 11:04:09 crc kubenswrapper[4733]: I0318 11:04:09.590310 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:04:09 crc kubenswrapper[4733]: E0318 11:04:09.590558 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:04:12 crc kubenswrapper[4733]: I0318 11:04:12.616766 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" exitCode=0 Mar 18 11:04:12 crc kubenswrapper[4733]: I0318 11:04:12.616860 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2"} Mar 18 11:04:12 crc kubenswrapper[4733]: I0318 11:04:12.617360 4733 scope.go:117] "RemoveContainer" containerID="bd29705735db3b754afb9922232f3ff6fa404f8d375f4f93c086696cc3583373" Mar 18 11:04:12 crc kubenswrapper[4733]: I0318 11:04:12.617990 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:04:12 crc kubenswrapper[4733]: E0318 11:04:12.618321 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:04:16 crc kubenswrapper[4733]: I0318 11:04:16.403890 4733 scope.go:117] "RemoveContainer" containerID="ae5d77ede52fa11bd913773d4add1f40cd6fcaf6154c4236eccd984879ea57ff" Mar 18 11:04:18 crc kubenswrapper[4733]: I0318 11:04:18.175428 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:04:18 crc kubenswrapper[4733]: E0318 11:04:18.176028 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:04:21 crc kubenswrapper[4733]: I0318 11:04:21.533740 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-zsljc_7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0/kube-rbac-proxy/0.log" Mar 18 11:04:21 crc kubenswrapper[4733]: I0318 11:04:21.587905 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-zsljc_7eed25d9-11cc-4ca1-b715-0a77d4dcc8e0/controller/0.log" Mar 18 11:04:21 crc kubenswrapper[4733]: I0318 11:04:21.769277 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-frr-files/0.log" Mar 18 11:04:21 crc kubenswrapper[4733]: I0318 11:04:21.936782 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-frr-files/0.log" Mar 18 11:04:21 crc kubenswrapper[4733]: I0318 11:04:21.990603 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-reloader/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.009971 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-metrics/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.017933 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-reloader/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.196730 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-reloader/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.230388 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-metrics/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.235123 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-frr-files/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.256286 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-metrics/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.417147 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/controller/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.450382 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-metrics/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.455242 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-reloader/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.460464 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/cp-frr-files/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.620578 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/frr-metrics/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.639892 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/kube-rbac-proxy-frr/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.656173 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/kube-rbac-proxy/0.log" Mar 18 11:04:22 crc kubenswrapper[4733]: I0318 11:04:22.997671 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-dr9dg_03476444-8ff8-4b1e-bcbc-ee654241370b/frr-k8s-webhook-server/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.009316 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/reloader/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.175153 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:04:23 crc kubenswrapper[4733]: E0318 11:04:23.175389 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.177395 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pc5zz_4a1a5332-5dcd-46ab-b584-0a8fb7feaa9e/frr/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.198035 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5ddc5ff65-jst9z_9731a250-9d78-43e0-bde3-7e769ea43d11/manager/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.336910 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c99d9f4d6-n5lc9_37ecdf54-7bcf-4d33-9cd9-f156974ea7f9/webhook-server/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.397673 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zg5cv_101c5687-bebd-449f-94c8-03077bf596d0/kube-rbac-proxy/0.log" Mar 18 11:04:23 crc kubenswrapper[4733]: I0318 11:04:23.637069 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zg5cv_101c5687-bebd-449f-94c8-03077bf596d0/speaker/0.log" Mar 18 11:04:26 crc kubenswrapper[4733]: I0318 11:04:26.176014 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:04:26 crc kubenswrapper[4733]: E0318 11:04:26.176566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:04:33 crc kubenswrapper[4733]: I0318 11:04:33.175841 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:04:33 crc kubenswrapper[4733]: E0318 11:04:33.176832 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:04:36 crc kubenswrapper[4733]: I0318 11:04:36.175590 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:04:36 crc kubenswrapper[4733]: E0318 11:04:36.176311 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:04:36 crc kubenswrapper[4733]: I0318 11:04:36.885702 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.087651 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.105656 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/pull/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.165045 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/pull/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.443326 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/pull/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.463499 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/extract/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.470851 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874rxg2p_3f95562e-ae03-4b2d-92b7-bc5593785f3c/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.628903 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.786598 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/pull/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.817364 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/pull/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.875462 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.974854 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/util/0.log" Mar 18 11:04:37 crc kubenswrapper[4733]: I0318 11:04:37.978050 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/extract/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.042597 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1d59fb_2d140fd7-f9d7-4432-aef0-ec0ab2e18cf6/pull/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.181539 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-utilities/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.327940 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-content/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.332324 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-utilities/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.351084 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-content/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.507414 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-utilities/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.541861 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/extract-content/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.765929 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-utilities/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.888608 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-content/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.906400 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c9s27_a3f55919-82b3-4117-8734-cb9a26364d83/registry-server/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.917973 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-utilities/0.log" Mar 18 11:04:38 crc kubenswrapper[4733]: I0318 11:04:38.941662 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-content/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.143630 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-utilities/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.147746 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/extract-content/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.353153 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z6qb2_8ae3847e-6357-46a1-9578-88deb6e1531b/marketplace-operator/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.452693 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-utilities/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.614373 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-smkxx_3b618f79-3791-49a8-a6aa-307fb25af727/registry-server/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.631407 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-content/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.654411 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-utilities/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.727308 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-content/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.857778 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-content/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.860915 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/extract-utilities/0.log" Mar 18 11:04:39 crc kubenswrapper[4733]: I0318 11:04:39.981453 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kcbhw_20640f37-bf35-4f24-abbb-b31cd00e5c9c/registry-server/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.037606 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-utilities/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.176800 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:04:40 crc kubenswrapper[4733]: E0318 11:04:40.176991 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.224415 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-content/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.240720 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-content/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.260082 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-utilities/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.423995 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-utilities/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.443507 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/extract-content/0.log" Mar 18 11:04:40 crc kubenswrapper[4733]: I0318 11:04:40.813242 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rs2b6_3018dd18-ee9f-44a1-ab22-a6bddde19b31/registry-server/0.log" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.496086 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:42 crc kubenswrapper[4733]: E0318 11:04:42.496923 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720459c9-dd9a-4d0f-8541-f4f2f578acc5" containerName="oc" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.496936 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="720459c9-dd9a-4d0f-8541-f4f2f578acc5" containerName="oc" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.497101 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="720459c9-dd9a-4d0f-8541-f4f2f578acc5" containerName="oc" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.498240 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.509769 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.546531 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.546785 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pddvl\" (UniqueName: \"kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.546860 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.647824 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.647888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.647980 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pddvl\" (UniqueName: \"kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.648458 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.648672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.674512 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pddvl\" (UniqueName: \"kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl\") pod \"certified-operators-w89tj\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:42 crc kubenswrapper[4733]: I0318 11:04:42.823496 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:43 crc kubenswrapper[4733]: I0318 11:04:43.373579 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:43 crc kubenswrapper[4733]: I0318 11:04:43.849246 4733 generic.go:334] "Generic (PLEG): container finished" podID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerID="d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0" exitCode=0 Mar 18 11:04:43 crc kubenswrapper[4733]: I0318 11:04:43.849334 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerDied","Data":"d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0"} Mar 18 11:04:43 crc kubenswrapper[4733]: I0318 11:04:43.849562 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerStarted","Data":"9af4ebac00ba06ba0a306bf985e9fada6fd3b57d11129aef1aab44ff8dd47136"} Mar 18 11:04:44 crc kubenswrapper[4733]: I0318 11:04:44.858523 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerStarted","Data":"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3"} Mar 18 11:04:45 crc kubenswrapper[4733]: I0318 11:04:45.178921 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:04:45 crc kubenswrapper[4733]: E0318 11:04:45.179172 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:04:45 crc kubenswrapper[4733]: I0318 11:04:45.867933 4733 generic.go:334] "Generic (PLEG): container finished" podID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerID="8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3" exitCode=0 Mar 18 11:04:45 crc kubenswrapper[4733]: I0318 11:04:45.867971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerDied","Data":"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3"} Mar 18 11:04:46 crc kubenswrapper[4733]: I0318 11:04:46.879311 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerStarted","Data":"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff"} Mar 18 11:04:46 crc kubenswrapper[4733]: I0318 11:04:46.908600 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w89tj" podStartSLOduration=2.465932184 podStartE2EDuration="4.908581473s" podCreationTimestamp="2026-03-18 11:04:42 +0000 UTC" firstStartedPulling="2026-03-18 11:04:43.850902584 +0000 UTC m=+3123.342636909" lastFinishedPulling="2026-03-18 11:04:46.293551873 +0000 UTC m=+3125.785286198" observedRunningTime="2026-03-18 11:04:46.903396627 +0000 UTC m=+3126.395130962" watchObservedRunningTime="2026-03-18 11:04:46.908581473 +0000 UTC m=+3126.400315798" Mar 18 11:04:47 crc kubenswrapper[4733]: I0318 11:04:47.175645 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:04:47 crc kubenswrapper[4733]: E0318 11:04:47.175885 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:04:52 crc kubenswrapper[4733]: I0318 11:04:52.824346 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:52 crc kubenswrapper[4733]: I0318 11:04:52.824878 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:52 crc kubenswrapper[4733]: I0318 11:04:52.878773 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:52 crc kubenswrapper[4733]: I0318 11:04:52.965680 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:53 crc kubenswrapper[4733]: I0318 11:04:53.136910 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:54 crc kubenswrapper[4733]: I0318 11:04:54.937869 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w89tj" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="registry-server" containerID="cri-o://43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff" gracePeriod=2 Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.175710 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:04:55 crc kubenswrapper[4733]: E0318 11:04:55.176433 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.375200 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.456320 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities\") pod \"720d76e8-3e14-45ec-87e6-99948b3e0f42\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.456421 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content\") pod \"720d76e8-3e14-45ec-87e6-99948b3e0f42\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.456490 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pddvl\" (UniqueName: \"kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl\") pod \"720d76e8-3e14-45ec-87e6-99948b3e0f42\" (UID: \"720d76e8-3e14-45ec-87e6-99948b3e0f42\") " Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.457349 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities" (OuterVolumeSpecName: "utilities") pod "720d76e8-3e14-45ec-87e6-99948b3e0f42" (UID: "720d76e8-3e14-45ec-87e6-99948b3e0f42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.463397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl" (OuterVolumeSpecName: "kube-api-access-pddvl") pod "720d76e8-3e14-45ec-87e6-99948b3e0f42" (UID: "720d76e8-3e14-45ec-87e6-99948b3e0f42"). InnerVolumeSpecName "kube-api-access-pddvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.560390 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pddvl\" (UniqueName: \"kubernetes.io/projected/720d76e8-3e14-45ec-87e6-99948b3e0f42-kube-api-access-pddvl\") on node \"crc\" DevicePath \"\"" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.560422 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.949750 4733 generic.go:334] "Generic (PLEG): container finished" podID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerID="43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff" exitCode=0 Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.949897 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerDied","Data":"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff"} Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.950028 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w89tj" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.950040 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w89tj" event={"ID":"720d76e8-3e14-45ec-87e6-99948b3e0f42","Type":"ContainerDied","Data":"9af4ebac00ba06ba0a306bf985e9fada6fd3b57d11129aef1aab44ff8dd47136"} Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.950063 4733 scope.go:117] "RemoveContainer" containerID="43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.970331 4733 scope.go:117] "RemoveContainer" containerID="8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.993078 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "720d76e8-3e14-45ec-87e6-99948b3e0f42" (UID: "720d76e8-3e14-45ec-87e6-99948b3e0f42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:04:55 crc kubenswrapper[4733]: I0318 11:04:55.993910 4733 scope.go:117] "RemoveContainer" containerID="d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.032323 4733 scope.go:117] "RemoveContainer" containerID="43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff" Mar 18 11:04:56 crc kubenswrapper[4733]: E0318 11:04:56.033394 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff\": container with ID starting with 43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff not found: ID does not exist" containerID="43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.033471 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff"} err="failed to get container status \"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff\": rpc error: code = NotFound desc = could not find container \"43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff\": container with ID starting with 43615e4c6b1987af73cae1c93ca995b55c2aafa63ad191baf1f760950b7c4bff not found: ID does not exist" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.033503 4733 scope.go:117] "RemoveContainer" containerID="8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3" Mar 18 11:04:56 crc kubenswrapper[4733]: E0318 11:04:56.033890 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3\": container with ID starting with 8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3 not found: ID does not exist" containerID="8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.033919 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3"} err="failed to get container status \"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3\": rpc error: code = NotFound desc = could not find container \"8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3\": container with ID starting with 8b6502424e46362b8ac33d9fd6c25d1864e353fe3a70af5033ba644217eb03c3 not found: ID does not exist" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.033941 4733 scope.go:117] "RemoveContainer" containerID="d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0" Mar 18 11:04:56 crc kubenswrapper[4733]: E0318 11:04:56.034368 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0\": container with ID starting with d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0 not found: ID does not exist" containerID="d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.034399 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0"} err="failed to get container status \"d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0\": rpc error: code = NotFound desc = could not find container \"d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0\": container with ID starting with d28f8f8f036e6e7300bc6459e017a6d89efcdd0467a2fe3dc572cd6efe822bf0 not found: ID does not exist" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.067953 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/720d76e8-3e14-45ec-87e6-99948b3e0f42-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.284886 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:56 crc kubenswrapper[4733]: I0318 11:04:56.296654 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w89tj"] Mar 18 11:04:57 crc kubenswrapper[4733]: I0318 11:04:57.175385 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:04:57 crc kubenswrapper[4733]: E0318 11:04:57.175743 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:04:57 crc kubenswrapper[4733]: I0318 11:04:57.191527 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" path="/var/lib/kubelet/pods/720d76e8-3e14-45ec-87e6-99948b3e0f42/volumes" Mar 18 11:05:00 crc kubenswrapper[4733]: I0318 11:05:00.175443 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:05:00 crc kubenswrapper[4733]: E0318 11:05:00.176316 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:05:09 crc kubenswrapper[4733]: I0318 11:05:09.175555 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:05:09 crc kubenswrapper[4733]: E0318 11:05:09.176425 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:05:10 crc kubenswrapper[4733]: I0318 11:05:10.175638 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:05:10 crc kubenswrapper[4733]: E0318 11:05:10.176077 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:05:12 crc kubenswrapper[4733]: I0318 11:05:12.175844 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:05:12 crc kubenswrapper[4733]: E0318 11:05:12.176984 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:05:22 crc kubenswrapper[4733]: I0318 11:05:22.175632 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:05:22 crc kubenswrapper[4733]: E0318 11:05:22.176402 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:05:23 crc kubenswrapper[4733]: I0318 11:05:23.175827 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:05:23 crc kubenswrapper[4733]: E0318 11:05:23.176330 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:05:26 crc kubenswrapper[4733]: I0318 11:05:26.176118 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:05:26 crc kubenswrapper[4733]: E0318 11:05:26.176795 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:05:35 crc kubenswrapper[4733]: I0318 11:05:35.176118 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:05:35 crc kubenswrapper[4733]: E0318 11:05:35.178386 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:05:38 crc kubenswrapper[4733]: I0318 11:05:38.176570 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:05:38 crc kubenswrapper[4733]: E0318 11:05:38.177276 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:05:41 crc kubenswrapper[4733]: I0318 11:05:41.180463 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:05:41 crc kubenswrapper[4733]: E0318 11:05:41.181002 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:05:49 crc kubenswrapper[4733]: I0318 11:05:49.176423 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:05:49 crc kubenswrapper[4733]: E0318 11:05:49.177159 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:05:52 crc kubenswrapper[4733]: I0318 11:05:52.175768 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:05:52 crc kubenswrapper[4733]: E0318 11:05:52.176566 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:05:54 crc kubenswrapper[4733]: I0318 11:05:54.175833 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:05:54 crc kubenswrapper[4733]: E0318 11:05:54.176377 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:05:57 crc kubenswrapper[4733]: I0318 11:05:57.482140 4733 generic.go:334] "Generic (PLEG): container finished" podID="5542a33f-3466-419c-af8f-3391bcc3d241" containerID="83201b6bc07225e6541cd95837440f81a991cdbcc76e2856eb4af65abc081fb9" exitCode=0 Mar 18 11:05:57 crc kubenswrapper[4733]: I0318 11:05:57.482244 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m754c/must-gather-dtbq6" event={"ID":"5542a33f-3466-419c-af8f-3391bcc3d241","Type":"ContainerDied","Data":"83201b6bc07225e6541cd95837440f81a991cdbcc76e2856eb4af65abc081fb9"} Mar 18 11:05:57 crc kubenswrapper[4733]: I0318 11:05:57.483512 4733 scope.go:117] "RemoveContainer" containerID="83201b6bc07225e6541cd95837440f81a991cdbcc76e2856eb4af65abc081fb9" Mar 18 11:05:57 crc kubenswrapper[4733]: I0318 11:05:57.849328 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m754c_must-gather-dtbq6_5542a33f-3466-419c-af8f-3391bcc3d241/gather/0.log" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.145246 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563866-78qk8"] Mar 18 11:06:00 crc kubenswrapper[4733]: E0318 11:06:00.145910 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="extract-utilities" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.145925 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="extract-utilities" Mar 18 11:06:00 crc kubenswrapper[4733]: E0318 11:06:00.145937 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="extract-content" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.145945 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="extract-content" Mar 18 11:06:00 crc kubenswrapper[4733]: E0318 11:06:00.145957 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="registry-server" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.145965 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="registry-server" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.146130 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="720d76e8-3e14-45ec-87e6-99948b3e0f42" containerName="registry-server" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.146718 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.148624 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.149197 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.150280 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.154583 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563866-78qk8"] Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.186142 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45g6m\" (UniqueName: \"kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m\") pod \"auto-csr-approver-29563866-78qk8\" (UID: \"259a6c5e-ad27-412a-b854-776f42d48c84\") " pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.287259 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45g6m\" (UniqueName: \"kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m\") pod \"auto-csr-approver-29563866-78qk8\" (UID: \"259a6c5e-ad27-412a-b854-776f42d48c84\") " pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.309054 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45g6m\" (UniqueName: \"kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m\") pod \"auto-csr-approver-29563866-78qk8\" (UID: \"259a6c5e-ad27-412a-b854-776f42d48c84\") " pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.466847 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:00 crc kubenswrapper[4733]: I0318 11:06:00.690317 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563866-78qk8"] Mar 18 11:06:01 crc kubenswrapper[4733]: I0318 11:06:01.511690 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563866-78qk8" event={"ID":"259a6c5e-ad27-412a-b854-776f42d48c84","Type":"ContainerStarted","Data":"0c1dd95218ba98497cb804ff0ac97f1c01408a8561346b9dcaac9335cfa062cd"} Mar 18 11:06:02 crc kubenswrapper[4733]: I0318 11:06:02.176379 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:06:02 crc kubenswrapper[4733]: E0318 11:06:02.176936 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:06:02 crc kubenswrapper[4733]: I0318 11:06:02.521559 4733 generic.go:334] "Generic (PLEG): container finished" podID="259a6c5e-ad27-412a-b854-776f42d48c84" containerID="3fa4661cb631b7da2b1c8775c9cfda48c185327f23a2376fd0ca5927a6b04720" exitCode=0 Mar 18 11:06:02 crc kubenswrapper[4733]: I0318 11:06:02.521632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563866-78qk8" event={"ID":"259a6c5e-ad27-412a-b854-776f42d48c84","Type":"ContainerDied","Data":"3fa4661cb631b7da2b1c8775c9cfda48c185327f23a2376fd0ca5927a6b04720"} Mar 18 11:06:03 crc kubenswrapper[4733]: I0318 11:06:03.857412 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.052211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45g6m\" (UniqueName: \"kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m\") pod \"259a6c5e-ad27-412a-b854-776f42d48c84\" (UID: \"259a6c5e-ad27-412a-b854-776f42d48c84\") " Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.059828 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m" (OuterVolumeSpecName: "kube-api-access-45g6m") pod "259a6c5e-ad27-412a-b854-776f42d48c84" (UID: "259a6c5e-ad27-412a-b854-776f42d48c84"). InnerVolumeSpecName "kube-api-access-45g6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.154331 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45g6m\" (UniqueName: \"kubernetes.io/projected/259a6c5e-ad27-412a-b854-776f42d48c84-kube-api-access-45g6m\") on node \"crc\" DevicePath \"\"" Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.543446 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563866-78qk8" event={"ID":"259a6c5e-ad27-412a-b854-776f42d48c84","Type":"ContainerDied","Data":"0c1dd95218ba98497cb804ff0ac97f1c01408a8561346b9dcaac9335cfa062cd"} Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.543492 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c1dd95218ba98497cb804ff0ac97f1c01408a8561346b9dcaac9335cfa062cd" Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.543530 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563866-78qk8" Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.916349 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563860-2k4dj"] Mar 18 11:06:04 crc kubenswrapper[4733]: I0318 11:06:04.921659 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563860-2k4dj"] Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.184963 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e41c70a-6d8d-47a8-9caf-57f46a60f96a" path="/var/lib/kubelet/pods/6e41c70a-6d8d-47a8-9caf-57f46a60f96a/volumes" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.188965 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m754c/must-gather-dtbq6"] Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.189734 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m754c/must-gather-dtbq6" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="copy" containerID="cri-o://1557b68a8a325a0ed8ba1781990482335f57ef77fd7aa971706340af195f6c7b" gracePeriod=2 Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.196765 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m754c/must-gather-dtbq6"] Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.553039 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m754c_must-gather-dtbq6_5542a33f-3466-419c-af8f-3391bcc3d241/copy/0.log" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.553850 4733 generic.go:334] "Generic (PLEG): container finished" podID="5542a33f-3466-419c-af8f-3391bcc3d241" containerID="1557b68a8a325a0ed8ba1781990482335f57ef77fd7aa971706340af195f6c7b" exitCode=143 Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.701601 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m754c_must-gather-dtbq6_5542a33f-3466-419c-af8f-3391bcc3d241/copy/0.log" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.705852 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.794922 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57p52\" (UniqueName: \"kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52\") pod \"5542a33f-3466-419c-af8f-3391bcc3d241\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.795009 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output\") pod \"5542a33f-3466-419c-af8f-3391bcc3d241\" (UID: \"5542a33f-3466-419c-af8f-3391bcc3d241\") " Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.810418 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52" (OuterVolumeSpecName: "kube-api-access-57p52") pod "5542a33f-3466-419c-af8f-3391bcc3d241" (UID: "5542a33f-3466-419c-af8f-3391bcc3d241"). InnerVolumeSpecName "kube-api-access-57p52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.906289 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57p52\" (UniqueName: \"kubernetes.io/projected/5542a33f-3466-419c-af8f-3391bcc3d241-kube-api-access-57p52\") on node \"crc\" DevicePath \"\"" Mar 18 11:06:05 crc kubenswrapper[4733]: I0318 11:06:05.977801 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5542a33f-3466-419c-af8f-3391bcc3d241" (UID: "5542a33f-3466-419c-af8f-3391bcc3d241"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.007562 4733 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5542a33f-3466-419c-af8f-3391bcc3d241-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.175906 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:06:06 crc kubenswrapper[4733]: E0318 11:06:06.176209 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.563893 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m754c_must-gather-dtbq6_5542a33f-3466-419c-af8f-3391bcc3d241/copy/0.log" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.564523 4733 scope.go:117] "RemoveContainer" containerID="1557b68a8a325a0ed8ba1781990482335f57ef77fd7aa971706340af195f6c7b" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.564543 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m754c/must-gather-dtbq6" Mar 18 11:06:06 crc kubenswrapper[4733]: I0318 11:06:06.582895 4733 scope.go:117] "RemoveContainer" containerID="83201b6bc07225e6541cd95837440f81a991cdbcc76e2856eb4af65abc081fb9" Mar 18 11:06:07 crc kubenswrapper[4733]: I0318 11:06:07.191781 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" path="/var/lib/kubelet/pods/5542a33f-3466-419c-af8f-3391bcc3d241/volumes" Mar 18 11:06:08 crc kubenswrapper[4733]: I0318 11:06:08.175591 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:06:08 crc kubenswrapper[4733]: E0318 11:06:08.176395 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:06:14 crc kubenswrapper[4733]: I0318 11:06:14.176280 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:06:14 crc kubenswrapper[4733]: E0318 11:06:14.177373 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:06:20 crc kubenswrapper[4733]: I0318 11:06:20.176280 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:06:20 crc kubenswrapper[4733]: E0318 11:06:20.176994 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:06:22 crc kubenswrapper[4733]: I0318 11:06:22.511524 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:06:22 crc kubenswrapper[4733]: E0318 11:06:22.512890 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:06:25 crc kubenswrapper[4733]: I0318 11:06:25.176048 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:06:25 crc kubenswrapper[4733]: E0318 11:06:25.177072 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:06:34 crc kubenswrapper[4733]: I0318 11:06:34.175867 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:06:34 crc kubenswrapper[4733]: E0318 11:06:34.177436 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:06:36 crc kubenswrapper[4733]: I0318 11:06:36.175633 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:06:36 crc kubenswrapper[4733]: E0318 11:06:36.176485 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:06:39 crc kubenswrapper[4733]: I0318 11:06:39.175687 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:06:39 crc kubenswrapper[4733]: E0318 11:06:39.176235 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:06:46 crc kubenswrapper[4733]: I0318 11:06:46.175882 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:06:46 crc kubenswrapper[4733]: E0318 11:06:46.176652 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:06:48 crc kubenswrapper[4733]: I0318 11:06:48.176086 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:06:48 crc kubenswrapper[4733]: E0318 11:06:48.176812 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:06:53 crc kubenswrapper[4733]: I0318 11:06:53.177379 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:06:53 crc kubenswrapper[4733]: E0318 11:06:53.178451 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:07:01 crc kubenswrapper[4733]: I0318 11:07:01.180423 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:07:01 crc kubenswrapper[4733]: I0318 11:07:01.181086 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:07:01 crc kubenswrapper[4733]: E0318 11:07:01.181179 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:07:01 crc kubenswrapper[4733]: E0318 11:07:01.181462 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:07:07 crc kubenswrapper[4733]: I0318 11:07:07.175140 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:07:07 crc kubenswrapper[4733]: E0318 11:07:07.175932 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:07:14 crc kubenswrapper[4733]: I0318 11:07:14.175076 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:07:14 crc kubenswrapper[4733]: E0318 11:07:14.175859 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:07:15 crc kubenswrapper[4733]: I0318 11:07:15.175805 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:07:15 crc kubenswrapper[4733]: E0318 11:07:15.176097 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:07:16 crc kubenswrapper[4733]: I0318 11:07:16.532922 4733 scope.go:117] "RemoveContainer" containerID="ee8a5931d088bb90e3f8edd41217a30f581b7d88c4f982136e16b0f2b145d28c" Mar 18 11:07:18 crc kubenswrapper[4733]: I0318 11:07:18.175936 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:07:18 crc kubenswrapper[4733]: E0318 11:07:18.176915 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:07:27 crc kubenswrapper[4733]: I0318 11:07:27.175526 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:07:27 crc kubenswrapper[4733]: I0318 11:07:27.176263 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:07:27 crc kubenswrapper[4733]: E0318 11:07:27.176593 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:07:27 crc kubenswrapper[4733]: E0318 11:07:27.176647 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:07:32 crc kubenswrapper[4733]: I0318 11:07:32.175709 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:07:32 crc kubenswrapper[4733]: E0318 11:07:32.176643 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:07:38 crc kubenswrapper[4733]: I0318 11:07:38.176122 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:07:38 crc kubenswrapper[4733]: E0318 11:07:38.176898 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:07:42 crc kubenswrapper[4733]: I0318 11:07:42.175303 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:07:42 crc kubenswrapper[4733]: E0318 11:07:42.176373 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:07:43 crc kubenswrapper[4733]: I0318 11:07:43.175940 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:07:43 crc kubenswrapper[4733]: E0318 11:07:43.176214 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2h7dp_openshift-machine-config-operator(6f75e1c5-e0c5-43df-944f-77b734070793)\"" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" podUID="6f75e1c5-e0c5-43df-944f-77b734070793" Mar 18 11:07:51 crc kubenswrapper[4733]: I0318 11:07:51.183487 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:07:51 crc kubenswrapper[4733]: E0318 11:07:51.184141 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:07:54 crc kubenswrapper[4733]: I0318 11:07:54.175533 4733 scope.go:117] "RemoveContainer" containerID="6b9a340729099e48708bb3e49a96ed003cdb26d857ad4f772c65d5062fdefcf9" Mar 18 11:07:55 crc kubenswrapper[4733]: I0318 11:07:55.416706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2h7dp" event={"ID":"6f75e1c5-e0c5-43df-944f-77b734070793","Type":"ContainerStarted","Data":"cfc089247a62ba1977b8f1dbc94d8aa7156f75d1768073acced9e580d06b9f83"} Mar 18 11:07:57 crc kubenswrapper[4733]: I0318 11:07:57.175178 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:07:57 crc kubenswrapper[4733]: E0318 11:07:57.175510 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.188656 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563868-bsjmk"] Mar 18 11:08:00 crc kubenswrapper[4733]: E0318 11:08:00.189659 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="259a6c5e-ad27-412a-b854-776f42d48c84" containerName="oc" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189676 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="259a6c5e-ad27-412a-b854-776f42d48c84" containerName="oc" Mar 18 11:08:00 crc kubenswrapper[4733]: E0318 11:08:00.189693 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="gather" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189701 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="gather" Mar 18 11:08:00 crc kubenswrapper[4733]: E0318 11:08:00.189714 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="copy" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189723 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="copy" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189937 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="copy" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189958 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5542a33f-3466-419c-af8f-3391bcc3d241" containerName="gather" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.189976 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="259a6c5e-ad27-412a-b854-776f42d48c84" containerName="oc" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.190662 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.192211 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.192310 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wmd5k" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.192993 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.197588 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563868-bsjmk"] Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.334949 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvqk9\" (UniqueName: \"kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9\") pod \"auto-csr-approver-29563868-bsjmk\" (UID: \"81e9d3af-d753-42e5-b7e3-529203e7cfab\") " pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.436660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvqk9\" (UniqueName: \"kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9\") pod \"auto-csr-approver-29563868-bsjmk\" (UID: \"81e9d3af-d753-42e5-b7e3-529203e7cfab\") " pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.462535 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvqk9\" (UniqueName: \"kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9\") pod \"auto-csr-approver-29563868-bsjmk\" (UID: \"81e9d3af-d753-42e5-b7e3-529203e7cfab\") " pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.511640 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.916736 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563868-bsjmk"] Mar 18 11:08:00 crc kubenswrapper[4733]: W0318 11:08:00.922230 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e9d3af_d753_42e5_b7e3_529203e7cfab.slice/crio-68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8 WatchSource:0}: Error finding container 68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8: Status 404 returned error can't find the container with id 68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8 Mar 18 11:08:00 crc kubenswrapper[4733]: I0318 11:08:00.925146 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 11:08:01 crc kubenswrapper[4733]: I0318 11:08:01.470350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" event={"ID":"81e9d3af-d753-42e5-b7e3-529203e7cfab","Type":"ContainerStarted","Data":"68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8"} Mar 18 11:08:02 crc kubenswrapper[4733]: I0318 11:08:02.481033 4733 generic.go:334] "Generic (PLEG): container finished" podID="81e9d3af-d753-42e5-b7e3-529203e7cfab" containerID="29bf663b381a7fa830f7e7010a5bc46c4197d3720124ddb060ebae43c856ab2e" exitCode=0 Mar 18 11:08:02 crc kubenswrapper[4733]: I0318 11:08:02.481095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" event={"ID":"81e9d3af-d753-42e5-b7e3-529203e7cfab","Type":"ContainerDied","Data":"29bf663b381a7fa830f7e7010a5bc46c4197d3720124ddb060ebae43c856ab2e"} Mar 18 11:08:03 crc kubenswrapper[4733]: I0318 11:08:03.821170 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:03 crc kubenswrapper[4733]: I0318 11:08:03.995402 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvqk9\" (UniqueName: \"kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9\") pod \"81e9d3af-d753-42e5-b7e3-529203e7cfab\" (UID: \"81e9d3af-d753-42e5-b7e3-529203e7cfab\") " Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.000401 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9" (OuterVolumeSpecName: "kube-api-access-xvqk9") pod "81e9d3af-d753-42e5-b7e3-529203e7cfab" (UID: "81e9d3af-d753-42e5-b7e3-529203e7cfab"). InnerVolumeSpecName "kube-api-access-xvqk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.097154 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvqk9\" (UniqueName: \"kubernetes.io/projected/81e9d3af-d753-42e5-b7e3-529203e7cfab-kube-api-access-xvqk9\") on node \"crc\" DevicePath \"\"" Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.176150 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:08:04 crc kubenswrapper[4733]: E0318 11:08:04.177062 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.509174 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" event={"ID":"81e9d3af-d753-42e5-b7e3-529203e7cfab","Type":"ContainerDied","Data":"68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8"} Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.509240 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68224e65287acce10060c36b5fe003286c5f71fcd676f7c816e02786f315b5c8" Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.509257 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563868-bsjmk" Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.883999 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563862-vmt2n"] Mar 18 11:08:04 crc kubenswrapper[4733]: I0318 11:08:04.889529 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563862-vmt2n"] Mar 18 11:08:05 crc kubenswrapper[4733]: I0318 11:08:05.191336 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad06f0e3-f671-426a-b6d8-793e87745364" path="/var/lib/kubelet/pods/ad06f0e3-f671-426a-b6d8-793e87745364/volumes" Mar 18 11:08:09 crc kubenswrapper[4733]: I0318 11:08:09.176483 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:08:09 crc kubenswrapper[4733]: E0318 11:08:09.177315 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:08:15 crc kubenswrapper[4733]: I0318 11:08:15.176706 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:08:15 crc kubenswrapper[4733]: E0318 11:08:15.177844 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:08:16 crc kubenswrapper[4733]: I0318 11:08:16.599242 4733 scope.go:117] "RemoveContainer" containerID="8d69124510c80bc1e6643920b76340f2795eccc769fbdff3665f6edcc2793fe5" Mar 18 11:08:20 crc kubenswrapper[4733]: I0318 11:08:20.175044 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:08:20 crc kubenswrapper[4733]: E0318 11:08:20.175802 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:08:26 crc kubenswrapper[4733]: I0318 11:08:26.175992 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:08:26 crc kubenswrapper[4733]: E0318 11:08:26.176706 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:08:34 crc kubenswrapper[4733]: I0318 11:08:34.175942 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:08:34 crc kubenswrapper[4733]: E0318 11:08:34.177586 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:08:40 crc kubenswrapper[4733]: I0318 11:08:40.175796 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:08:40 crc kubenswrapper[4733]: E0318 11:08:40.176514 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:08:49 crc kubenswrapper[4733]: I0318 11:08:49.176165 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:08:49 crc kubenswrapper[4733]: E0318 11:08:49.176959 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:08:52 crc kubenswrapper[4733]: I0318 11:08:52.175672 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:08:52 crc kubenswrapper[4733]: E0318 11:08:52.176363 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:09:00 crc kubenswrapper[4733]: I0318 11:09:00.917526 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:00 crc kubenswrapper[4733]: E0318 11:09:00.918426 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e9d3af-d753-42e5-b7e3-529203e7cfab" containerName="oc" Mar 18 11:09:00 crc kubenswrapper[4733]: I0318 11:09:00.918457 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e9d3af-d753-42e5-b7e3-529203e7cfab" containerName="oc" Mar 18 11:09:00 crc kubenswrapper[4733]: I0318 11:09:00.918661 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e9d3af-d753-42e5-b7e3-529203e7cfab" containerName="oc" Mar 18 11:09:00 crc kubenswrapper[4733]: I0318 11:09:00.920005 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:00 crc kubenswrapper[4733]: I0318 11:09:00.933086 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.090354 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wswjh\" (UniqueName: \"kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.090409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.090526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.191961 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wswjh\" (UniqueName: \"kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.192011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.192092 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.192694 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.192694 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.210247 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wswjh\" (UniqueName: \"kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh\") pod \"community-operators-qjhkq\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.246785 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:01 crc kubenswrapper[4733]: I0318 11:09:01.527867 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:02 crc kubenswrapper[4733]: I0318 11:09:02.007521 4733 generic.go:334] "Generic (PLEG): container finished" podID="d5decccb-25db-497f-834d-00a802b2a5ca" containerID="b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b" exitCode=0 Mar 18 11:09:02 crc kubenswrapper[4733]: I0318 11:09:02.007564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerDied","Data":"b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b"} Mar 18 11:09:02 crc kubenswrapper[4733]: I0318 11:09:02.007868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerStarted","Data":"e9440f409036519ec55c7c9ca7bffba341f56cdd2766643a8f9051792924a14b"} Mar 18 11:09:03 crc kubenswrapper[4733]: I0318 11:09:03.176968 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:09:03 crc kubenswrapper[4733]: E0318 11:09:03.177441 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:09:04 crc kubenswrapper[4733]: I0318 11:09:04.024495 4733 generic.go:334] "Generic (PLEG): container finished" podID="d5decccb-25db-497f-834d-00a802b2a5ca" containerID="0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3" exitCode=0 Mar 18 11:09:04 crc kubenswrapper[4733]: I0318 11:09:04.024564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerDied","Data":"0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3"} Mar 18 11:09:04 crc kubenswrapper[4733]: I0318 11:09:04.176081 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:09:04 crc kubenswrapper[4733]: E0318 11:09:04.176713 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:09:05 crc kubenswrapper[4733]: I0318 11:09:05.032863 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerStarted","Data":"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3"} Mar 18 11:09:05 crc kubenswrapper[4733]: I0318 11:09:05.052607 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qjhkq" podStartSLOduration=2.505431591 podStartE2EDuration="5.052584947s" podCreationTimestamp="2026-03-18 11:09:00 +0000 UTC" firstStartedPulling="2026-03-18 11:09:02.014469089 +0000 UTC m=+3381.506203434" lastFinishedPulling="2026-03-18 11:09:04.561622435 +0000 UTC m=+3384.053356790" observedRunningTime="2026-03-18 11:09:05.046794004 +0000 UTC m=+3384.538528359" watchObservedRunningTime="2026-03-18 11:09:05.052584947 +0000 UTC m=+3384.544319272" Mar 18 11:09:11 crc kubenswrapper[4733]: I0318 11:09:11.246931 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:11 crc kubenswrapper[4733]: I0318 11:09:11.247509 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:11 crc kubenswrapper[4733]: I0318 11:09:11.315393 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:12 crc kubenswrapper[4733]: I0318 11:09:12.151245 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:12 crc kubenswrapper[4733]: I0318 11:09:12.201765 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.123275 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qjhkq" podUID="d5decccb-25db-497f-834d-00a802b2a5ca" containerName="registry-server" containerID="cri-o://7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3" gracePeriod=2 Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.550130 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.686427 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wswjh\" (UniqueName: \"kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh\") pod \"d5decccb-25db-497f-834d-00a802b2a5ca\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.686514 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content\") pod \"d5decccb-25db-497f-834d-00a802b2a5ca\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.686702 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities\") pod \"d5decccb-25db-497f-834d-00a802b2a5ca\" (UID: \"d5decccb-25db-497f-834d-00a802b2a5ca\") " Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.687654 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities" (OuterVolumeSpecName: "utilities") pod "d5decccb-25db-497f-834d-00a802b2a5ca" (UID: "d5decccb-25db-497f-834d-00a802b2a5ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.694583 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh" (OuterVolumeSpecName: "kube-api-access-wswjh") pod "d5decccb-25db-497f-834d-00a802b2a5ca" (UID: "d5decccb-25db-497f-834d-00a802b2a5ca"). InnerVolumeSpecName "kube-api-access-wswjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.788600 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 11:09:14 crc kubenswrapper[4733]: I0318 11:09:14.788647 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wswjh\" (UniqueName: \"kubernetes.io/projected/d5decccb-25db-497f-834d-00a802b2a5ca-kube-api-access-wswjh\") on node \"crc\" DevicePath \"\"" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.133693 4733 generic.go:334] "Generic (PLEG): container finished" podID="d5decccb-25db-497f-834d-00a802b2a5ca" containerID="7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3" exitCode=0 Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.133825 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjhkq" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.133747 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerDied","Data":"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3"} Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.133985 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjhkq" event={"ID":"d5decccb-25db-497f-834d-00a802b2a5ca","Type":"ContainerDied","Data":"e9440f409036519ec55c7c9ca7bffba341f56cdd2766643a8f9051792924a14b"} Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.134030 4733 scope.go:117] "RemoveContainer" containerID="7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.159283 4733 scope.go:117] "RemoveContainer" containerID="0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.198765 4733 scope.go:117] "RemoveContainer" containerID="b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.241040 4733 scope.go:117] "RemoveContainer" containerID="7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3" Mar 18 11:09:15 crc kubenswrapper[4733]: E0318 11:09:15.241529 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3\": container with ID starting with 7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3 not found: ID does not exist" containerID="7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.241657 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3"} err="failed to get container status \"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3\": rpc error: code = NotFound desc = could not find container \"7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3\": container with ID starting with 7298bb974b60eb967ef1ec7369e4539b9ca038a55f17d333b8184b4668a6a4a3 not found: ID does not exist" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.241755 4733 scope.go:117] "RemoveContainer" containerID="0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3" Mar 18 11:09:15 crc kubenswrapper[4733]: E0318 11:09:15.242250 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3\": container with ID starting with 0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3 not found: ID does not exist" containerID="0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.242282 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3"} err="failed to get container status \"0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3\": rpc error: code = NotFound desc = could not find container \"0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3\": container with ID starting with 0b138f51f2df78a11e054926beb21b5f8ff84a8876455818cb6870ff0aa34fc3 not found: ID does not exist" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.242302 4733 scope.go:117] "RemoveContainer" containerID="b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b" Mar 18 11:09:15 crc kubenswrapper[4733]: E0318 11:09:15.242570 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b\": container with ID starting with b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b not found: ID does not exist" containerID="b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.242672 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b"} err="failed to get container status \"b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b\": rpc error: code = NotFound desc = could not find container \"b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b\": container with ID starting with b84fc31514ed603bba303ec28f5a426ecb6ae7c046b4a75d643b055090d6437b not found: ID does not exist" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.409078 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5decccb-25db-497f-834d-00a802b2a5ca" (UID: "d5decccb-25db-497f-834d-00a802b2a5ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.482847 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.489039 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qjhkq"] Mar 18 11:09:15 crc kubenswrapper[4733]: I0318 11:09:15.499905 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5decccb-25db-497f-834d-00a802b2a5ca-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 11:09:16 crc kubenswrapper[4733]: I0318 11:09:16.176037 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01" Mar 18 11:09:16 crc kubenswrapper[4733]: I0318 11:09:16.176305 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:09:17 crc kubenswrapper[4733]: I0318 11:09:17.160725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerStarted","Data":"42cbdac80d1cd5197063da49b635b07d62e4adb2d6aae581ef46d16897659830"} Mar 18 11:09:17 crc kubenswrapper[4733]: I0318 11:09:17.161002 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 11:09:17 crc kubenswrapper[4733]: I0318 11:09:17.163438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerStarted","Data":"438887752c2380d0a118f44c0a43f524012895d063bcabcf9aab6778e9825f97"} Mar 18 11:09:17 crc kubenswrapper[4733]: I0318 11:09:17.163663 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 11:09:17 crc kubenswrapper[4733]: I0318 11:09:17.212105 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5decccb-25db-497f-834d-00a802b2a5ca" path="/var/lib/kubelet/pods/d5decccb-25db-497f-834d-00a802b2a5ca/volumes" Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.207284 4733 generic.go:334] "Generic (PLEG): container finished" podID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" containerID="42cbdac80d1cd5197063da49b635b07d62e4adb2d6aae581ef46d16897659830" exitCode=0 Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.207438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b4a4e3e2-bd4d-4f8d-97bc-51267378ab03","Type":"ContainerDied","Data":"42cbdac80d1cd5197063da49b635b07d62e4adb2d6aae581ef46d16897659830"} Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.207944 4733 scope.go:117] "RemoveContainer" containerID="309174b794edb6ce74f0fdb4a12ba0a1a8e65a9dcfd1acde2e49c6c8caf177d2" Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.208827 4733 scope.go:117] "RemoveContainer" containerID="42cbdac80d1cd5197063da49b635b07d62e4adb2d6aae581ef46d16897659830" Mar 18 11:09:21 crc kubenswrapper[4733]: E0318 11:09:21.209270 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-cell1-server-0_openstack(b4a4e3e2-bd4d-4f8d-97bc-51267378ab03)\"" pod="openstack/rabbitmq-cell1-server-0" podUID="b4a4e3e2-bd4d-4f8d-97bc-51267378ab03" Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.215317 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0570ce4-1455-4698-85cf-01f7108d9e7f" containerID="438887752c2380d0a118f44c0a43f524012895d063bcabcf9aab6778e9825f97" exitCode=0 Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.215360 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f0570ce4-1455-4698-85cf-01f7108d9e7f","Type":"ContainerDied","Data":"438887752c2380d0a118f44c0a43f524012895d063bcabcf9aab6778e9825f97"} Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.215939 4733 scope.go:117] "RemoveContainer" containerID="438887752c2380d0a118f44c0a43f524012895d063bcabcf9aab6778e9825f97" Mar 18 11:09:21 crc kubenswrapper[4733]: E0318 11:09:21.216138 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"rabbitmq\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=rabbitmq pod=rabbitmq-server-0_openstack(f0570ce4-1455-4698-85cf-01f7108d9e7f)\"" pod="openstack/rabbitmq-server-0" podUID="f0570ce4-1455-4698-85cf-01f7108d9e7f" Mar 18 11:09:21 crc kubenswrapper[4733]: I0318 11:09:21.775825 4733 scope.go:117] "RemoveContainer" containerID="32438c6a9409b79313e3b8972bb637f88d330bfefddb767cde843ec5e6f0eb01"